• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 17:32
CEST 23:32
KST 06:32
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
BGE Stara Zagora 2025: Info & Preview17Code S RO12 Preview: GuMiho, Bunny, SHIN, ByuN3The Memories We Share - Facing the Final(?) GSL46Code S RO12 Preview: Cure, Zoun, Solar, Creator4[ASL19] Finals Preview: Daunting Task30
Community News
GSL Ro4 and Finals moved to Sunday June 15th10Weekly Cups (May 27-June 1): ByuN goes back-to-back0EWC 2025 Regional Qualifier Results26Code S RO12 Results + RO8 Groups (2025 Season 2)3Weekly Cups (May 19-25): Hindsight is 20/20?0
StarCraft 2
General
Jim claims he and Firefly were involved in match-fixing BGE Stara Zagora 2025: Info & Preview The Memories We Share - Facing the Final(?) GSL GSL Ro4 and Finals moved to Sunday June 15th Serious Question: Mech
Tourneys
Bellum Gens Elite: Stara Zagora 2025 $25,000+ WardiTV 2025 Series Sparkling Tuna Cup - Weekly Open Tournament SOOP Starcraft Global #21 $5,100+ SEL Season 2 Championship (SC: Evo)
Strategy
[G] Darkgrid Layout Simple Questions Simple Answers [G] PvT Cheese: 13 Gate Proxy Robo
Custom Maps
[UMS] Zillion Zerglings
External Content
Mutation # 476 Charnel House Mutation # 475 Hard Target Mutation # 474 Futile Resistance Mutation # 473 Cold is the Void
Brood War
General
Will foreigners ever be able to challenge Koreans? FlaSh Witnesses SCV Pull Off the Impossible vs Shu BW General Discussion BGH auto balance -> http://bghmmr.eu/ Battle.net is not working
Tourneys
[ASL19] Grand Finals [Megathread] Daily Proleagues Small VOD Thread 2.0 [BSL20] GosuLeague RO16 - Tue & Wed 20:00+CET
Strategy
I am doing this better than progamers do. [G] How to get started on ladder as a new Z player
Other Games
General Games
Monster Hunter Wilds Stormgate/Frost Giant Megathread Nintendo Switch Thread Path of Exile Mechabellum
Dota 2
Official 'what is Dota anymore' discussion
League of Legends
LiquidLegends to reintegrate into TL.net
Heroes of the Storm
Simple Questions, Simple Answers
Hearthstone
Heroes of StarCraft mini-set
TL Mafia
Vanilla Mini Mafia TL Mafia Community Thread TL Mafia Plays: Diplomacy TL Mafia: Generative Agents Showdown Survivor II: The Amazon
Community
General
US Politics Mega-thread Russo-Ukrainian War Thread Things Aren’t Peaceful in Palestine Vape Nation Thread European Politico-economics QA Mega-thread
Fan Clubs
Maru Fan Club Serral Fan Club
Media & Entertainment
Korean Music Discussion [Manga] One Piece
Sports
2024 - 2025 Football Thread Formula 1 Discussion NHL Playoffs 2024
World Cup 2022
Tech Support
Computer Build, Upgrade & Buying Resource Thread Cleaning My Mechanical Keyboard
TL Community
The Automated Ban List
Blogs
Heero Yuy & the Tax…
KrillinFromwales
Research study on team perfo…
TrAiDoS
I was completely wrong ab…
jameswatts
Need Your Help/Advice
Glider
Trip to the Zoo
micronesia
Poker
Nebuchad
Info SLEgma_12
SLEgma_12
Customize Sidebar...

Website Feedback

Closed Threads



Active: 18273 users

Critique my abstract (and...

Blogs > EatThePath
Post a Reply
1 2 Next All
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
Last Edited: 2011-07-15 20:09:06
July 14 2011 20:34 GMT
#1
...find out what I did this summer in 300 words or less)

Update

Hi TL, I need to write an abstract (as though for a scientific publication) as a preliminary step in documenting the work I've done in a research internship this summer.

The work I did was actually not too complicated, based on the software aspect of forensic DNA analysis.

I would like you to give me feedback on whether or not the following makes sense, and whether it provides an adequate summary. Of course you don't know the details, so this is an outside looking in sort of test! If you feel like you have a question about something because I left it out, let me know. If you feel like you get it, insofar as your knowledge of DNA forensics takes you, that's good, even if that's not very far.

Also I'm worried about the structure and the flow, not so much whether the jargon is comprehensible. I know it's hard to sort that out completely. This isn't meant to be for laymen, but it should be more accessible than a document meant just for specialists. If you have any kind of science background, most of it should make a bit of sense.

Abstract

Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. A forensic analyst typically reviews the results of the software analysis, which depends on user-defined parameters in conjunction with rules that handle variations. For example, a commonly used peak detection threshold for heterozygous loci is 100RFUs, whereas at homozygous loci the threshold is 200RFUs. The analyst makes allele calls (or confirms those made by the software) using information provided by the software which is the result of involved computations. In the case of the example, a peak height below the detection threshold is deemed statistically undependable. This peak height can vary based on what parameters are used, and it also depends on aspects of the signal processing algorithms that cannot be altered by the user. While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


Thanks for your time. =)

*
Comprehensive strategic intention: DNE
Servius_Fulvius
Profile Joined August 2009
United States947 Posts
July 14 2011 20:52 GMT
#2
Feel free to ignore my take on it. I'm a chemical engineer, but each scientific discipline has its own standards when it comes to literature. My overall opinion is that you included too much introductory information. Each paper has a section set aside for the introduction and it seems like you'd be repeating yourself a lot there (much to the chagrin of those actually reading the whole thing). If I were writing this it would look something like:

On July 15 2011 05:34 EatThePath wrote:
Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


That's just a crude cut and paste. Since I'm not familiar with the subject I can't tell you if I cut out something really important. Then again, a lot of that material seemed like something someone knowledgable in the field would already know. Again, it's great to have in the body of the paper, but if those reading it are anything like my graduate adviser who tells us to read the abstract and conclusion and only then decide if it's worth out time, then they may not have the patience.

It's a good write-up, though. I have not composed a paper yet, but within the next year I'll be in your shoes!
GreatFall
Profile Blog Joined January 2010
United States1061 Posts
Last Edited: 2011-07-14 21:04:35
July 14 2011 21:02 GMT
#3
I have a Ph.D. in Genetics. This is how I would word the genetic parts of your abstract.

"Modern forensic DNA analysis relies on software programs to interpret and display data obtained from electrophoresis of DNA fragments."

rest looks good
Inventor of the 'Burning Tide' technique to quickly getting Outmatched Crusher achivement :D
Probe1
Profile Blog Joined August 2010
United States17920 Posts
July 14 2011 21:41 GMT
#4
I know what you did last summer.


Just kidding I don't know enough to give a serious comment. Hopefully I made you smile though.
우정호 KT_VIOLET 1988 - 2012 While we are postponing, life speeds by
spacefarm
Profile Joined July 2010
United States112 Posts
Last Edited: 2011-07-14 23:17:30
July 14 2011 23:16 GMT
#5
I can't comment on content, but I believe it is a little too long. If you go to google scholar and read random journal articles you will see that abstracts are typically a little shorter and I think some of the content in the first paragraph can be included in the introduction portion of your paper.

good job on doing summer research though, PhD programs <3 that.

edit: gak, what servius said ^^
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 14 2011 23:48 GMT
#6
On July 15 2011 05:52 Servius_Fulvius wrote:+ Show Spoiler +

Feel free to ignore my take on it. I'm a chemical engineer, but each scientific discipline has its own standards when it comes to literature. My overall opinion is that you included too much introductory information. Each paper has a section set aside for the introduction and it seems like you'd be repeating yourself a lot there (much to the chagrin of those actually reading the whole thing). If I were writing this it would look something like:

On July 15 2011 05:34 EatThePath wrote:
Modern forensic DNA typing relies on software programs to interpret and display data obtained from capillary electrophoresis of DNA fragments. Common functions performed by the software include baselining of multiple signal channels, noise-filtering, and smoothing, which are preliminary to analytical functions such as identification of artifacts and allele calling. This study investigated the disparities in peak heights given by four software programs, Genemapper IDX (Applied Biosystems), Genemarker HID (who), FSSi3 (who), and TrueAllele (who), operating on the same raw data. We found highly correlated differences in peak heights between these programs. Using the same analysis parameters, this led to variations in the profile from a single sample as determined by the different programs when the profile contained alleles near the peak detection threshold. We conducted a simulation to test the efficacy of adjustments based on regression analysis of our peak height data. Using these adjustments, we were able to significantly reduce discrepancies in profiles exhibiting near-threshold peaks.


That's just a crude cut and paste. Since I'm not familiar with the subject I can't tell you if I cut out something really important. Then again, a lot of that material seemed like something someone knowledgable in the field would already know. Again, it's great to have in the body of the paper, but if those reading it are anything like my graduate adviser who tells us to read the abstract and conclusion and only then decide if it's worth out time, then they may not have the patience.

It's a good write-up, though. I have not composed a paper yet, but within the next year I'll be in your shoes!



In my job I read plenty of scientific abstracts, mainly cancer research. I agree with this post wholeheartedly. The abstract is intended to give the reader a rundown of what to expect when reading your paper. I would cut the intro section completely and focus on your second paragraph. Normally you want to outline
  • what you did
  • what you expected
  • what you got
  • why this is significant
  • and what it means for future research.

If the person reading doesn't know about DNA typing then you can outline it in the introduction section of the paper.

In short keep your audience in the front of your mind. They are most likely experts who want new information, they probably do not want to be taught the basics.

As for the writing style, wording and flow. It seems fine to me. The flow is good, and the wording not overly specific. All in all a nicely written abstract, just refocus the content a little.

Good luck mate
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
Malgrif
Profile Blog Joined March 2010
Canada1095 Posts
July 15 2011 00:37 GMT
#7
the first paragraph is useless for an abstract. you want to tell the reader what the paper contains, not background information as why the stuff it contains is important
for there to be pro there has to be noob.
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 02:19 GMT
#8
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 02:39 GMT
#9
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 03:24 GMT
#10
On July 15 2011 11:39 Probulous wrote:
Show nested quote +
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.
Comprehensive strategic intention: DNE
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 03:24 GMT
#11
On July 15 2011 06:41 Probe1 wrote:
I know what you did last summer.


Just kidding I don't know enough to give a serious comment. Hopefully I made you smile though.


You did, hehe.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 03:38 GMT
#12
On July 15 2011 12:24 EatThePath wrote:
Show nested quote +
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
Show nested quote +
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 04:10 GMT
#13
On July 15 2011 12:38 Probulous wrote:
Show nested quote +
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
Last Edited: 2011-07-15 04:25:26
July 15 2011 04:21 GMT
#14
On July 15 2011 13:10 EatThePath wrote:
Show nested quote +
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.

Edit 2: Turns out you have been here way longer than I have, so now I look like an idiot
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
Last Edited: 2011-07-15 04:30:33
July 15 2011 04:28 GMT
#15
On July 15 2011 13:21 Probulous wrote:
Show nested quote +
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 04:34 GMT
#16
On July 15 2011 13:28 EatThePath wrote:
Show nested quote +
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 04:45 GMT
#17
On July 15 2011 13:34 Probulous wrote:
Show nested quote +
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 05:03 GMT
#18
On July 15 2011 13:45 EatThePath wrote:
Show nested quote +
On July 15 2011 13:34 Probulous wrote:
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?


I didn't quite mean it like that. Just that some people design their life around what they want to be.
Anyway, Kurszweil is a very interesting read because he is quite logical and clear about what is currently happening. He then takes things to their logical conclusion. Nothing more. His premise is that with the continued exponential increase in computing power and developments in AI we will eventually create a machine that passes the turing test. He predicts somewhere around 2040 based on current rates of development.

Haven't read much about information as a physical entity, I suppose that makes sense in that information can be seen as the transfer of states. If all information that I understand is captured in the structure of my brain then it must have a physical structure? Interesting...

My job is way off base. I work for a pharma company looking after our preclinical research projects. Basically get to read scientists abstracts before they go to print. Pretty awesome in that respect.

Blog series sounds like a plan. It is kinda hard to organise stuff being in Oz.
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
EatThePath
Profile Blog Joined September 2009
United States3943 Posts
July 15 2011 05:16 GMT
#19
On July 15 2011 14:03 Probulous wrote:
Show nested quote +
On July 15 2011 13:45 EatThePath wrote:
On July 15 2011 13:34 Probulous wrote:
On July 15 2011 13:28 EatThePath wrote:
On July 15 2011 13:21 Probulous wrote:
On July 15 2011 13:10 EatThePath wrote:
On July 15 2011 12:38 Probulous wrote:
On July 15 2011 12:24 EatThePath wrote:
On July 15 2011 11:39 Probulous wrote:
On July 15 2011 11:19 EatThePath wrote:
Thanks guys for the comments! Keep 'em coming if any more people happen to see this. ^^

I totally agree with all of you about the intro content. Yuck. This abstract won't actually be used for a paper, it's meant to introduce / summarize a poster (like you see in the halls at a lab building) that will be presented to other students/researchers of disparate backgrounds.

Personally I think it's pretty stupid but that's why I have all that intro junk in there, because the abstract is supposed to stand on its own in providing context and significance.

I'm glad it comes across okay. I'll definitely revise the second section to provide more "going forward from these results" comments.

Your input is invaluable folks. ;D


Well in that case you will need some introduction stuff.

just a couple of sentences outlining what the background of DNA typing is. You want people to read your content, not the background. Are you giving a presentation with this or it will stand on its own?


I'm not entirely sure what the people running the program have in mind. I know that I'm giving a presentation to other people here with me (30-40 other interns at this school) where we share our research with each other but.. I really couldn't care less about that. I think that's mostly for practice, eh. They so far have "encouraged us strongly" to submit our research to student conferences that pander to this sort of thing. (Yay, more padding for graduate school aps). And I understand that many of these conferences weed you out initially based on abstract alone. Perhaps they also want to use the abstracts for some sort of review / wider publication survey of research kind of thing.

That's as much as I know about that.

To complicate matters, in my own personal case I'm also helping to write the actual paper, which will have different verbage than this and my poster. So I might be overemphasizing the differences between "the real one" and my "sharing one".

Also, to make the point of the study clear, I have to make this clear:
While validation of the most commonly used software is extensive, it focuses on parameter settings that reproduce correct results, but does not otherwise account for the underlying signal processing.

And in order for that to mean anything to a wider audience I have to explain about the software a bit.

I don't want to bother you guys with the nitty gritty reasoning of revising this to death. I'll go into it more if you want.


Well this complicates things I guess. Still I would suggest using the paper abstract as your base and the simplify for the student audience. If these conferences are going to be used for your post-grad entry you want to be seen as scientific first and foremost. You're pretty luck to be an author on a paper so early. That's a big bonus.

As for the specifics /details. My degree was in bioinformatics so this is right up my alley When you say signal processing, you mean the translation of the raw signals into base pairs? If so, interesting area of comparison, I guess this is a neglected area of validation. Using the peak heights is interesting too. Were any of your difference significant enough to warrant a rethink of the software itself.

I mean if you are getting different sequences with different software packages, how do you interpet your results...


Oh hai you know what I'm talking about. ;D

Yes I'm super excited to be an author! I'll be revising with my mentor / graduate student who is helping me too, so I think with all this assistance I can strike the right balance. I am thinking now to err on the side of straight and to-the-point.

So... I am dealing with STR profiling where you essentially count up the repeats to by measuring fragment length and assign an allele number. To do this you run a size standard to sync up BP size to time in the run, and you also run a ladder which has every allele to provide you the software with "bins" so it knows what allele number to call based on where the peak shows up.

The peak heights don't matter at all most of the time. If they are super high it can cause artifacts and if they are unexpectedly low you probably had a problem during your PCR or whatnot. Also, if you have low copy to start with or degraded DNA (like resumed remains) you'll have low peaks cause you just don't have that much to start with so it's low signal. So you get situations where your peaks are too low and they don't meet the detection threshold. In this case you throw out that locus as inconclusive which gives you a partial profile. In analogy, you have a partial fingerprint.

Almost everyone in the US uses the same piece of software from the company whose instruments almost everyone uses too. But there are new programs entering the market and they are not being looked at a whole lot yet. So my study found that different programs have slightly higher or lower peak heights overall, which is the result of slightly different approaches in the initial raw data processing. They all give you the same profile under normal circumstances because that just depends on the sizing the peaks, aka fragment length (not height Y, just X position). When you have low peaks (for whatever reason) you can have a situation where in one program, it is just under the threshold so you don't include that locus, and in another program it meets the threshold and you include the locus. So you have differing completeness of profiles which are concordant, but one has more information. A given locus can represent discriminatory power (likelihood of matching a random human) of like 1/1000. If you gain or lose two loci your odds you tell the jury in court can differ by a million-fold.

That's a dramatic way to put it but it illustrates the point--you want concordance in your profiling. We tested the solution of just adjusting the threshold based on our regressions. Which works pretty well, it helps a lot. But at a certain level your signal processing is just not the same and there's no way around that. Big picture, it's not a huge deal but it's something that people should think about.

Wow I gassed on. Hope you find it interesting.


Yeah that was what I expected. One or two loci isn't going to break the bank unless you are really low on your initial sample, in which case there are other sources of error. It is an interesting area to look at, simply because most people just assume that what they read is correct.

Well done on the work. Good luck with the paper, any ideas what you want to do after this?

Edit: Oh and you will soon learn that on TL, there is always someone who knows what you are talking about. Doesn't matter the subject. There are some serious brains on this site And no I don't consider myself a brain.


Exactly, people just go with the program. To be fair the demand for profiling (not just in a criminal forensic application) is high and growing, so there will be lots of people who are basically technicians. In that kind of world we should make sure all the automated stuff works properly first. And in my experience the forensic community is utterly dedicated to that, even if they are pulled and pressured from different angles. Like state mandated caseload processing rates.

What do I want to do? I'm going to make artificial intelligences like Cortana from Halo.

edit. lol no worries I barely consider myself a TLer. I have seen some impressive folks though, did you see that thread "what is your occupation / aspiration?" I was blown away.


Yup, makes my career seems really mundane. Oh well, my job was never going to define me.

You are certainly reaching for the stars. I don't play halo, so I am afraid that reference just flew by me.

Gvien your direction I would imagine you've read a bit of Kurzweil's theory on the singularity. You want to be a part of developing that? It is pretty mind-bending, not sure I could give up my inferior fleshiness.


Nah man, there's every chance I'll end up "not being defined by my career" but that's what I'm shooting for. Finally. After 8 years without a university degree I find my calling the last few months. I haven't heard of Kurzweil before but a quick googling shows me things I can jive with. I just finished reading "The User Illusion" by Tor Norretranders which synthesises basically everything I hold dear that I've come to know about math science and the universe. It's a thesis about consciousness, but it's based on physics, ultimately.

I don't go in for the sci fi motivation "lets build the future" kind of mentality even though I'm partial to it, very. For me it's more about understanding the universe and where that takes us as humans. The salient point there is that information is a physical quantity. The ramifications of that are pretty intense. I would love to tell you about it at length, lol. ;D

Maybe there will be a blog series. Or a skype science club meeting when it's not midnight.

What are you doing with a bioinformatics degree that has you reading abstracts?


I didn't quite mean it like that. Just that some people design their life around what they want to be.
Anyway, Kurszweil is a very interesting read because he is quite logical and clear about what is currently happening. He then takes things to their logical conclusion. Nothing more. His premise is that with the continued exponential increase in computing power and developments in AI we will eventually create a machine that passes the turing test. He predicts somewhere around 2040 based on current rates of development.

Haven't read much about information as a physical entity, I suppose that makes sense in that information can be seen as the transfer of states. If all information that I understand is captured in the structure of my brain then it must have a physical structure? Interesting...

My job is way off base. I work for a pharma company looking after our preclinical research projects. Basically get to read scientists abstracts before they go to print. Pretty awesome in that respect.

Blog series sounds like a plan. It is kinda hard to organise stuff being in Oz.


Sounds like a good setup for keeping the brain engaged and fending off boredom. Maybe I'm naive but I'm always stunned when I hear about niches like yours, "you get payed for that??".

I just read all up and down the wikipedia page. So did you read the book? I've given a lot of thought to the general social problem of increasing rates of technological development. The singularity notion is new to me and it's cool. I need to read more before I decide for sure to debate it. The part I bolded, I have beef with that. For next time.

Cheers, thanks again for the tips, I really appreciate it. ^^
Comprehensive strategic intention: DNE
Probulous
Profile Blog Joined March 2011
Australia3894 Posts
July 15 2011 05:24 GMT
#20
No Sweat mate, PM next time you post a blog. I am not on all the time
"Dude has some really interesting midgame switches that I wouldn't have expected. "I violated your house" into "HIHO THE DAIRY OH!" really threw me. You don't usually expect children's poetry harass as a follow up " - AmericanUmlaut
1 2 Next All
Please log in or register to reply.
Live Events Refresh
Next event in 12h 28m
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
UpATreeSC 163
ForJumy 101
StarCraft: Brood War
MaD[AoV]61
Dota 2
LuMiX2
Counter-Strike
fl0m6737
Foxcn731
flusha252
Stewie2K125
Super Smash Bros
C9.Mang0642
hungrybox278
Mew2King68
Heroes of the Storm
Grubby3592
Liquid`Hasu507
Other Games
summit1g6618
FrodaN1221
elazer343
Pyrionflax97
Trikslyr77
ZombieGrub63
ViBE61
Tefel4
Organizations
Dota 2
PGL Dota 2 - Secondary Stream1123
Other Games
BasetradeTV27
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
sctven
[ Show 21 non-featured ]
StarCraft 2
• davetesta44
• musti20045 28
• Kozan
• sooper7s
• Migwel
• AfreecaTV YouTube
• LaughNgamezSOOP
• intothetv
• IndyKCrew
StarCraft: Brood War
• blackmanpl 44
• Eskiya23 15
• FirePhoenix3
• STPLYoutube
• ZZZeroYoutube
• BSLYoutube
Dota 2
• masondota21316
League of Legends
• Doublelift5042
• TFBlade1378
• Shiphtur517
Other Games
• imaqtpie1360
• Scarra595
Upcoming Events
The PondCast
12h 28m
Bellum Gens Elite
13h 28m
WardiTV Invitational
13h 28m
Replay Cast
1d 2h
OSC
1d 2h
Bellum Gens Elite
1d 13h
WardiTV Invitational
1d 16h
BSL 2v2 ProLeague
1d 21h
Replay Cast
2 days
CranKy Ducklings
2 days
[ Show More ]
SC Evo League
2 days
Bellum Gens Elite
2 days
Fire Grow Cup
2 days
CSO Contender
2 days
BSL: ProLeague
2 days
StRyKeR vs MadiNho
Cross vs UltrA
TT1 vs JDConan
Bonyth vs Sziky
Replay Cast
3 days
SOOP Global
3 days
Creator vs Rogue
Cure vs Classic
SOOP
3 days
SHIN vs GuMiho
Sparkling Tuna Cup
3 days
AllThingsProtoss
3 days
Fire Grow Cup
3 days
BSL: ProLeague
3 days
HBO vs Doodle
spx vs Tech
DragOn vs Hawk
Dewalt vs TerrOr
Replay Cast
4 days
Replay Cast
5 days
Replay Cast
5 days
WardiTV Invitational
5 days
GSL Code S
6 days
Rogue vs GuMiho
Maru vs Solar
Liquipedia Results

Completed

CSL Season 17: Qualifier 1
DreamHack Dallas 2025
Heroes 10 EU

Ongoing

JPL Season 2
BSL 2v2 Season 3
BSL Season 20
KCM Race Survival 2025 Season 2
NPSL S3
Rose Open S1
CSL Season 17: Qualifier 2
2025 GSL S2
Bellum Gens Elite Stara Zagora 2025
BLAST.tv Austin Major 2025
ESL Impact League Season 7
IEM Dallas 2025
PGL Astana 2025
Asian Champions League '25
ECL Season 49: Europe
BLAST Rivals Spring 2025
MESA Nomadic Masters
CCT Season 2 Global Finals
IEM Melbourne 2025
YaLLa Compass Qatar 2025
PGL Bucharest 2025
BLAST Open Spring 2025

Upcoming

CSL 17: 2025 SUMMER
Copa Latinoamericana 4
CSLPRO Last Chance 2025
CSLAN 2025
K-Championship
SEL Season 2 Championship
Esports World Cup 2025
HSC XXVII
Championship of Russia 2025
Murky Cup #2
BLAST Bounty Fall 2025
BLAST Bounty Fall Qual
IEM Cologne 2025
FISSURE Playground #1
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2025 TLnet. All Rights Reserved.