• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 08:46
CEST 14:46
KST 21:46
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
BGE Stara Zagora 2025: Info & Preview24Code S RO12 Preview: GuMiho, Bunny, SHIN, ByuN3The Memories We Share - Facing the Final(?) GSL46Code S RO12 Preview: Cure, Zoun, Solar, Creator4[ASL19] Finals Preview: Daunting Task30
Community News
[BSL20] ProLeague: Bracket Stage & Dates6GSL Ro4 and Finals moved to Sunday June 15th12Weekly Cups (May 27-June 1): ByuN goes back-to-back0EWC 2025 Regional Qualifier Results26Code S RO12 Results + RO8 Groups (2025 Season 2)3
StarCraft 2
General
The SCII GOAT: A statistical Evaluation Magnus Carlsen and Fabi review Clem's chess game. BGE Stara Zagora 2025: Info & Preview Buy driving license without exam for over 41 count Jim claims he and Firefly were involved in match-fixing
Tourneys
Bellum Gens Elite: Stara Zagora 2025 SOOPer7s Showmatches 2025 Cheeseadelphia 2025 - Open Bracket LAN! $25,000+ WardiTV 2025 Series Sparkling Tuna Cup - Weekly Open Tournament
Strategy
[G] Darkgrid Layout Simple Questions Simple Answers [G] PvT Cheese: 13 Gate Proxy Robo
Custom Maps
[UMS] Zillion Zerglings
External Content
Mutation # 476 Charnel House Mutation # 475 Hard Target Mutation # 474 Futile Resistance Mutation # 473 Cold is the Void
Brood War
General
BW General Discussion Will foreigners ever be able to challenge Koreans? BGH auto balance -> http://bghmmr.eu/ [BSL20] ProLeague: Bracket Stage & Dates I made an ASL quiz
Tourneys
[BSL20] ProLeague Bracket Stage - Day 2 [BSL20] ProLeague Bracket Stage - Day 1 [BSL 2v2] ProLeague Season 3 - Friday 21:00 CET [ASL19] Grand Finals
Strategy
I am doing this better than progamers do. [G] How to get started on ladder as a new Z player
Other Games
General Games
Nintendo Switch Thread Path of Exile Stormgate/Frost Giant Megathread Mechabellum Monster Hunter Wilds
Dota 2
Official 'what is Dota anymore' discussion
League of Legends
LiquidLegends to reintegrate into TL.net
Heroes of the Storm
Heroes of the Storm 2.0 Simple Questions, Simple Answers
Hearthstone
Heroes of StarCraft mini-set
TL Mafia
TL Mafia Community Thread Vanilla Mini Mafia
Community
General
US Politics Mega-thread Russo-Ukrainian War Thread Things Aren’t Peaceful in Palestine Vape Nation Thread European Politico-economics QA Mega-thread
Fan Clubs
Maru Fan Club Serral Fan Club
Media & Entertainment
Korean Music Discussion [Manga] One Piece
Sports
2024 - 2025 Football Thread Formula 1 Discussion NHL Playoffs 2024
World Cup 2022
Tech Support
Computer Build, Upgrade & Buying Resource Thread Cleaning My Mechanical Keyboard
TL Community
The Automated Ban List
Blogs
Cognitive styles x game perf…
TrAiDoS
StarCraft improvement
iopq
Heero Yuy & the Tax…
KrillinFromwales
I was completely wrong ab…
jameswatts
Need Your Help/Advice
Glider
Trip to the Zoo
micronesia
Poker
Nebuchad
Customize Sidebar...

Website Feedback

Closed Threads



Active: 20543 users

Critique my abstract (and...

Blogs > EatThePath
Post a Reply
Normal
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
Last Edited: 2011-07-15 20:09:06
July 14 2011 20:34 GMT
#1
...find out what I did this summer in 300 words or less)

Update

Hi TL, I need to write an abstract (as though for a scientific publication) as a preliminary step in documenting the work I've done in a research internship this summer.

The work I did was actually not too complicated, based on the software aspect of forensic DNA analysis.

I would like you to give me feedback on whether or not the following makes sense, and whether it provides an adequate summary. Of course you don't know the details, so this is an outside looking in sort of test! If you feel like you have a question about something because I left it out, let me know. If you feel like you get it, insofar as your knowledge of DNA forensics takes you, that's good, even if that's not very far.

Also I'm worried about the structure and the flow, not so much whether the jargon is comprehensible. I know it's hard to sort that out completely. This isn't meant to be for laymen, but it should be more accessible than a document meant just for specialists. If you have any kind of science background, most of it should make a bit of sense.

Abstract

Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. A forensic analyst typically reviews the results of the software analysis, which depends on user-defined parameters in conjunction with rules that handle variations. For example, a commonly used peak detection threshold for heterozygous loci is 100RFUs, whereas at homozygous loci the threshold is 200RFUs. The analyst makes allele calls (or confirms those made by the software) using information provided by the software which is the result of involved computations. In the case of the example, a peak height below the detection threshold is deemed statistically undependable. This peak height can vary based on what parameters are used, and it also depends on aspects of the signal processing algorithms that cannot be altered by the user. While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


Thanks for your time. =)

*
Comprehensive strategic intention: DNE
Servius_Fulvius
Profile Joined August 2009
United States947 Posts
July 14 2011 20:52 GMT
#2
Feel free to ignore my take on it. I'm a chemical engineer, but each scientific discipline has its own standards when it comes to literature. My overall opinion is that you included too much introductory information. Each paper has a section set aside for the introduction and it seems like you'd be repeating yourself a lot there (much to the chagrin of those actually reading the whole thing). If I were writing this it would look something like:

On July 15 2011 05:34 EatThePath wrote:
Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


That's just a crude cut and paste. Since I'm not familiar with the subject I can't tell you if I cut out something really important. Then again, a lot of that material seemed like something someone knowledgable in the field would already know. Again, it's great to have in the body of the paper, but if those reading it are anything like my graduate adviser who tells us to read the abstract and conclusion and only then decide if it's worth out time, then they may not have the patience.

It's a good write-up, though. I have not composed a paper yet, but within the next year I'll be in your shoes!
GreatFall
Profile Blog Joined January 2010
United States1061 Posts
Last Edited: 2011-07-14 21:04:35
July 14 2011 21:02 GMT
#3
I have a Ph.D. in Genetics. This is how I would word the genetic parts of your abstract.

"Modern forensic DNA analysis relies on software programs to interpret and display data obtained from electrophoresis of DNA fragments."

rest looks good
Inventor of the 'Burning Tide' technique to quickly getting Outmatched Crusher achivement :D
Probe1
Profile Blog Joined August 2010
United States17920 Posts
July 14 2011 21:41 GMT
#4
I know what you did last summer.


Just kidding I don't know enough to give a serious comment. Hopefully I made you smile though.
우정호 KT_VIOLET 1988 - 2012 While we are postponing, life speeds by
spacefarm
Profile Joined July 2010
United States112 Posts
Last Edited: 2011-07-14 23:17:30
July 14 2011 23:16 GMT
#5
I can't comment on content, but I believe it is a little too long. If you go to google scholar and read random journal articles you will see that abstracts are typically a little shorter and I think some of the content in the first paragraph can be included in the introduction portion of your paper.

good job on doing summer research though, PhD programs <3 that.

edit: gak, what servius said ^^
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 14 2011 23:48 GMT
#6
On July 15 2011 05:52 Servius_Fulvius wrote:+ Show Spoiler +

Feel free to ignore my take on it. I'm a chemical engineer, but each scientific discipline has its own standards when it comes to literature. My overall opinion is that you included too much introductory information. Each paper has a section set aside for the introduction and it seems like you'd be repeating yourself a lot there (much to the chagrin of those actually reading the whole thing). If I were writing this it would look something like:

On July 15 2011 05:34 EatThePath wrote:
Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


That's just a crude cut and paste. Since I'm not familiar with the subject I can't tell you if I cut out something really important. Then again, a lot of that material seemed like something someone knowledgable in the field would already know. Again, it's great to have in the body of the paper, but if those reading it are anything like my graduate adviser who tells us to read the abstract and conclusion and only then decide if it's worth out time, then they may not have the patience.

It's a good write-up, though. I have not composed a paper yet, but within the next year I'll be in your shoes!



In my job I read plenty of scientific abstracts, mainly cancer research. I agree with this post wholeheartedly. The abstract is intended to give the reader a rundown of what to expect when reading your paper. I would cut the intro section completely and focus on your second paragraph. Normally you want to outline
  • what you did
  • what you expected
  • what you got
  • why this is significant
  • and what it means for future research.

If the person reading doesn't know about DNA typing then you can outline it in the introduction section of the paper.

In short keep your audience in the front of your mind. They are most likely experts who want new information, they probably do not want to be taught the basics.

As for the writing style, wording and flow. It seems fine to me. The flow is good, and the wording not overly specific. All in all a nicely written abstract, just refocus the content a little.

Good luck mate
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
Malgrif
Profile Blog Joined March 2010
Canada1095 Posts
July 15 2011 00:37 GMT
#7
the first paragraph is useless for an abstract. you want to tell the reader what the paper contains, not background information as why the stuff it contains is important
for there to be pro there has to be noob.
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 02:19 GMT
#8
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 02:39 GMT
#9
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 03:24 GMT
#10
On July 15 2011 11:39 Probulous wrote:
Show nested quote +
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.
Comprehensive strategic intention: DNE
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 03:24 GMT
#11
On July 15 2011 06:41 Probe1 wrote:
I know what you did last summer.


Just kidding I don't know enough to give a serious comment. Hopefully I made you smile though.


You did, hehe.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 03:38 GMT
#12
On July 15 2011 12:24 EatThePath wrote:
Show nested quote +
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
Show nested quote +
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 04:10 GMT
#13
On July 15 2011 12:38 Probulous wrote:
Show nested quote +
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
Last Edited: 2011-07-15 04:25:26
July 15 2011 04:21 GMT
#14
On July 15 2011 13:10 EatThePath wrote:
Show nested quote +
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.

Edit 2: Turns out you have been here way longer than I have, so now I look like an idiot
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
Last Edited: 2011-07-15 04:30:33
July 15 2011 04:28 GMT
#15
On July 15 2011 13:21 Probulous wrote:
Show nested quote +
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 04:34 GMT
#16
On July 15 2011 13:28 EatThePath wrote:
Show nested quote +
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 04:45 GMT
#17
On July 15 2011 13:34 Probulous wrote:
Show nested quote +
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 05:03 GMT
#18
On July 15 2011 13:45 EatThePath wrote:
Show nested quote +
On July 15 2011 13:34 Probulous wrote:
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?


I didn't quite mean it like that. Just that some people design their life around what they want to be.
Anyway, Kurszweil is a very interesting read because he is quite logical and clear about what is currently happening. He then takes things to their logical conclusion. Nothing more. His premise is that with the continued exponential increase in computing power and developments in AI we will eventually create a machine that passes the turing test. He predicts somewhere around 2040 based on current rates of development.

Haven't read much about information as a physical entity, I suppose that makes sense in that information can be seen as the transfer of states. If all information that I understand is captured in the structure of my brain then it must have a physical structure? Interesting...

My job is way off base. I work for a pharma company looking after our preclinical research projects. Basically get to read scientists abstracts before they go to print. Pretty awesome in that respect.

Blog series sounds like a plan. It is kinda hard to organise stuff being in Oz.
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 05:16 GMT
#19
On July 15 2011 14:03 Probulous wrote:
Show nested quote +
On July 15 2011 13:45 EatThePath wrote:
On July 15 2011 13:34 Probulous wrote:
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?


I didn't quite mean it like that. Just that some people design their life around what they want to be.
Anyway, Kurszweil is a very interesting read because he is quite logical and clear about what is currently happening. He then takes things to their logical conclusion. Nothing more. His premise is that with the continued exponential increase in computing power and developments in AI we will eventually create a machine that passes the turing test. He predicts somewhere around 2040 based on current rates of development.

Haven't read much about information as a physical entity, I suppose that makes sense in that information can be seen as the transfer of states. If all information that I understand is captured in the structure of my brain then it must have a physical structure? Interesting...

My job is way off base. I work for a pharma company looking after our preclinical research projects. Basically get to read scientists abstracts before they go to print. Pretty awesome in that respect.

Blog series sounds like a plan. It is kinda hard to organise stuff being in Oz.


Sounds like a good setup for keeping the brain engaged and fending off boredom. Maybe I'm naive but I'm always stunned when I hear about niches like yours, "you get payed for that??".

I just read all up and down the wikipedia page. So did you read the book? I've given a lot of thought to the general social problem of increasing rates of technological development. The singularity notion is new to me and it's cool. I need to read more before I decide for sure to debate it. The part I bolded, I have beef with that. For next time.

Cheers, thanks again for the tips, I really appreciate it. ^^
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 05:24 GMT
#20
No Sweat mate, PM next time you post a blog. I am not on all the time
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 20:07 GMT
#21
Not that anyone cares, but this is the final version:

+ Show Spoiler +

Forensic DNA typing relies on software programs to process and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by these software programs include baselining, color separation, and smoothing, which precede analytical functions such as peak detection, sizing, and allele calling. The performance of many of these functions can be altered by changing user-defined parameters, which are determined by a laboratory’s validation studies. For example, a commonly used analytical peak detection threshold for allele calls is 100RFU (relative fluorescence units). The analyst confirms allele calls resulting from the involved computations performed by the software. In the case of the example, a peak height below the detection threshold is deemed statistically undependable, and the allele information it represents would not be reported in the DNA profile. Therefore, peak height is an important factor when interpreting DNA profiles. Because each program has its own algorithms, the peak height can vary based on differences in raw data processing.

This study investigated the disparities in peak heights produced by four software programs, Genemapper® ID-X (Applied Biosystems, Foster City, CA), GeneMarker® HID (Soft Genetics, State College, PA), FSS-i3™ (Promega Corporation, Madison, WI), and TrueAllele® Casework (CyberGenetics, Pittsburgh, PA), operating on the same raw data. We found correlated differences in peak heights between these programs. Variations in the reportable profile from a single sample exhibiting alleles near the peak detection threshold were observed. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.
Comprehensive strategic intention: DNE
Normal
Please log in or register to reply.
Live Events Refresh
WardiTV Invitational
11:00
WardiTV June Group B
Creator vs Jumy
ByuN vs Cure
Cure vs Jumy
ByuN vs Creator
Creator vs Cure
ByuN vs Jumy
WardiTV401
IntoTheiNu 10
LiquipediaDiscussion
Bellum Gens Elite
10:00
Stara Zagora 2025 Day 3
Serral vs ZounLIVE!
SKillous vs Krystianer
Harstem vs Lambo
Reynor vs Clem
ShoWTimE vs Ryung
Bellum Gens Elite2915
ComeBackTV 1024
TaKeTV 494
IndyStarCraft 298
Rex209
3DClanTV 181
LiquipediaDiscussion
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
Bellum Gens Elite2915
Lowko348
IndyStarCraft 298
Hui .223
Rex 209
CosmosSc2 81
EnDerr 4
StarCraft: Brood War
Britney 61444
Hyuk 2034
Shuttle 1433
Light 1348
Jaedong 766
EffOrt 688
hero 668
ToSsGirL 456
BeSt 332
Stork 268
[ Show more ]
Mini 267
ZerO 179
TY 69
soO 42
sorry 40
[sc1f]eonzerg 39
Sea.KH 37
sas.Sziky 34
Mind 31
JYJ29
sSak 23
Barracks 20
yabsab 17
Backho 16
HiyA 14
Sacsri 13
SilentControl 12
ajuk12(nOOB) 12
GoRush 8
Noble 5
Dota 2
XcaliburYe733
Fuzer 275
Counter-Strike
flusha266
markeloff120
Other Games
singsing2394
B2W.Neo856
DeMusliM435
crisheroes367
elazer214
Mew2King177
ArmadaUGS133
XaKoH 74
QueenE27
ZerO(Twitch)11
Organizations
Dota 2
PGL Dota 2 - Main Stream4777
Other Games
BasetradeTV33
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 13 non-featured ]
StarCraft 2
• StrangeGG 56
• AfreecaTV YouTube
• intothetv
• Kozan
• IndyKCrew
• LaughNgamezSOOP
• Migwel
• sooper7s
StarCraft: Brood War
• BSLYoutube
• STPLYoutube
• ZZZeroYoutube
League of Legends
• Nemesis6509
• Jankos2091
Upcoming Events
BSL 2v2 ProLeague
6h 15m
Replay Cast
11h 15m
CranKy Ducklings
21h 15m
SC Evo League
23h 15m
Bellum Gens Elite
23h 45m
Fire Grow Cup
1d 2h
CSO Contender
1d 4h
BSL: ProLeague
1d 5h
StRyKeR vs MadiNho
Cross vs UltrA
TT1 vs JDConan
Bonyth vs Sziky
Replay Cast
1d 11h
SOOP Global
1d 14h
Creator vs Rogue
Cure vs Classic
[ Show More ]
SOOP
1d 20h
Classic vs GuMiho
Sparkling Tuna Cup
1d 21h
AllThingsProtoss
1d 22h
Fire Grow Cup
2 days
BSL: ProLeague
2 days
HBO vs Doodle
spx vs Tech
DragOn vs Hawk
Dewalt vs TerrOr
Replay Cast
2 days
Replay Cast
3 days
Replay Cast
3 days
WardiTV Invitational
3 days
WardiTV Invitational
3 days
GSL Code S
4 days
Rogue vs GuMiho
Maru vs Solar
Replay Cast
5 days
GSL Code S
5 days
herO vs TBD
Classic vs TBD
The PondCast
5 days
Replay Cast
6 days
GSL Code S
6 days
WardiTV Invitational
6 days
Liquipedia Results

Completed

CSL Season 17: Qualifier 1
DreamHack Dallas 2025
Heroes 10 EU

Ongoing

JPL Season 2
BSL 2v2 Season 3
BSL Season 20
KCM Race Survival 2025 Season 2
NPSL S3
Rose Open S1
CSL Season 17: Qualifier 2
2025 GSL S2
BGE Stara Zagora 2025
BLAST.tv Austin Major 2025
ESL Impact League Season 7
IEM Dallas 2025
PGL Astana 2025
Asian Champions League '25
ECL Season 49: Europe
BLAST Rivals Spring 2025
MESA Nomadic Masters
CCT Season 2 Global Finals
IEM Melbourne 2025
YaLLa Compass Qatar 2025
PGL Bucharest 2025
BLAST Open Spring 2025

Upcoming

CSL 17: 2025 SUMMER
Copa Latinoamericana 4
CSLPRO Last Chance 2025
CSLPRO Chat StarLAN 3
K-Championship
SEL Season 2 Championship
Esports World Cup 2025
HSC XXVII
Championship of Russia 2025
Murky Cup #2
BLAST Bounty Fall 2025
BLAST Bounty Fall Qual
IEM Cologne 2025
FISSURE Playground #1
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2025 TLnet. All Rights Reserved.