• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 04:24
CEST 10:24
KST 17:24
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
Code S Season 1 (2026) - RO4 & Finals Preview4[ASL21] Ro4 Preview: On Course12Code S Season 1 - RO8 Preview7[ASL21] Ro8 Preview Pt2: Progenitors8Code S Season 1 - RO12 Group A: Rogue, Percival, Solar, Zoun13
Community News
Code S Season 1 (2026) - RO8 Results2Weekly Cups (May 4-10): Clem, MaxPax, herO win1Maestros of The Game 2 announcement and schedule !11Weekly Cups (April 27-May 4): Clem takes triple0RSL Revival: Season 5 - Qualifiers and Main Event12
StarCraft 2
General
Code S Season 1 (2026) - RO4 & Finals Preview Code S Season 1 (2026) - RO8 Results Code S Season 1 (2026) - RO12 Results Team Liquid Map Contest #22 - The Finalists MaNa leaves Team Liquid
Tourneys
GSL Code S Season 1 (2026) Sparkling Tuna Cup - Weekly Open Tournament KSL Week 89 2026 GSL Season 2 Qualifiers Maestros of The Game 2 announcement and schedule !
Strategy
Custom Maps
[D]RTS in all its shapes and glory <3 [A] Nemrods 1/4 players
External Content
Mutation # 525 Wheel of Misfortune The PondCast: SC2 News & Results Mutation # 524 Death and Taxes Mutation # 523 Firewall
Brood War
General
Pros React to: TvT Masterclass in FlaSh vs Light vespene.gg — BW replays in browser BGH Auto Balance -> http://bghmmr.eu/ BW General Discussion ASL21 General Discussion
Tourneys
[BSL22] RO8 Bracket Stage + Another TieBreaker [ASL21] Semifinals B [ASL21] Ro8 Day 4 Escore Tournament StarCraft Season 2
Strategy
Muta micro map competition Fighting Spirit mining rates [G] Hydra ZvZ: An Introduction Simple Questions, Simple Answers
Other Games
General Games
Path of Exile Stormgate/Frost Giant Megathread Nintendo Switch Thread Warcraft III: The Frozen Throne Starcraft Tabletop Miniature Game
Dota 2
The Story of Wings Gaming
League of Legends
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Deck construction bug Heroes of StarCraft mini-set
TL Mafia
Vanilla Mini Mafia Mafia Game Mode Feedback/Ideas TL Mafia Community Thread Five o'clock TL Mafia
Community
General
US Politics Mega-thread YouTube Thread European Politico-economics QA Mega-thread Russo-Ukrainian War Thread UK Politics Mega-thread
Fan Clubs
The herO Fan Club!
Media & Entertainment
[Manga] One Piece Anime Discussion Thread [Req][Books] Good Fantasy/SciFi books
Sports
2024 - 2026 Football Thread McBoner: A hockey love story Formula 1 Discussion
World Cup 2022
Tech Support
streaming software Strange computer issues (software) [G] How to Block Livestream Ads
TL Community
The Automated Ban List
Blogs
How EEG Data Can Predict Gam…
TrAiDoS
ramps on octagon
StaticNine
Funny Nicknames
LUCKY_NOOB
Customize Sidebar...

Website Feedback

Closed Threads



Active: 2034 users

US Politics Mega-thread - Page 2958

Forum Index > Closed
Post a Reply
Prev 1 2956 2957 2958 2959 2960 10093 Next
Read the rules in the OP before posting, please.

In order to ensure that this thread continues to meet TL standards and follows the proper guidelines, we will be enforcing the rules in the OP more strictly. Be sure to give them a re-read to refresh your memory! The vast majority of you are contributing in a healthy way, keep it up!

NOTE: When providing a source, explain why you feel it is relevant and what purpose it adds to the discussion if it's not obvious.
Also take note that unsubstantiated tweets/posts meant only to rekindle old arguments can result in a mod action.
oneofthem
Profile Blog Joined November 2005
Cayman Islands24199 Posts
February 18 2016 18:56 GMT
#59141
On February 19 2016 02:35 Gorsameth wrote:
Show nested quote +
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.
We have fed the heart on fantasies, the heart's grown brutal from the fare, more substance in our enmities than in our love
Plansix
Profile Blog Joined April 2011
United States60190 Posts
February 18 2016 18:56 GMT
#59142
On February 19 2016 03:48 Acrofales wrote:
Show nested quote +
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.

From a legal standpoint, that makes it very hard to prove a lot of things. If they can never open an Iphone for review, you could harass someone from it forever, sending death threats and other terrible thing and never be charged. As long as you did it all on the phone, including the creation of the address, it could be impossible to prove you sent the threats.

A tiny computer that can never be opened by anyone but the owner is the dream of almost every slightly smart criminal.
I have the Honor to be your Obedient Servant, P.6
TL+ Member
Soap
Profile Blog Joined April 2010
Brazil1546 Posts
Last Edited: 2016-02-18 19:10:45
February 18 2016 19:03 GMT
#59143
On February 19 2016 03:48 Acrofales wrote:
Show nested quote +
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.


They can seek the manufacturer assistance, who replied they'd have to make a key to all of their safes. That can be construed as a violation of the Fourth Amendment requirement that searches must be specific.
puerk
Profile Joined February 2015
Germany855 Posts
February 18 2016 19:08 GMT
#59144
On February 19 2016 03:56 oneofthem wrote:
Show nested quote +
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
WolfintheSheep
Profile Joined June 2011
Canada14127 Posts
February 18 2016 19:11 GMT
#59145
On February 19 2016 03:56 Plansix wrote:
Show nested quote +
On February 19 2016 03:48 Acrofales wrote:
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.

From a legal standpoint, that makes it very hard to prove a lot of things. If they can never open an Iphone for review, you could harass someone from it forever, sending death threats and other terrible thing and never be charged. As long as you did it all on the phone, including the creation of the address, it could be impossible to prove you sent the threats.

A tiny computer that can never be opened by anyone but the owner is the dream of almost every slightly smart criminal.

This exists already, and it's called the human brain.

Realistically, everything you send out on a phone still exists on the receiver's phone, and the records of some message being sent to that exact destination which align perfectly with the time stamps still exist. You don't need to break open a criminal's phone to prove that they were transmitting the incriminating messages beyond a reasonable doubt.

Honestly, people pretend that secrets didn't exist before computers, and that the situation is somehow worse. In reality it's far, far better for law enforcement, even with encryption, because you can't brute force or hack a person (legally) who is keeping secrets in their head. but a phone has no rights.
Average means I'm better than half of you.
oneofthem
Profile Blog Joined November 2005
Cayman Islands24199 Posts
February 18 2016 19:13 GMT
#59146
On February 19 2016 04:08 puerk wrote:
Show nested quote +
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?
We have fed the heart on fantasies, the heart's grown brutal from the fare, more substance in our enmities than in our love
puerk
Profile Joined February 2015
Germany855 Posts
February 18 2016 19:15 GMT
#59147
On February 19 2016 04:13 oneofthem wrote:
Show nested quote +
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

creating it after the fact only works through a vulnerability in the system, so doing it shows the existence of said vulnerability: in this case there seems to be a way to force a locked iphone to update its firmware to a faulty one, that will give access

if this is possible is still unclear, but if apple does it, the systems flaw is exposed and others will copy the technique
Gorsameth
Profile Joined April 2010
Netherlands22372 Posts
February 18 2016 19:17 GMT
#59148
On February 19 2016 04:13 oneofthem wrote:
Show nested quote +
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

How often do we need to lay out the path this takes before you understand why it is a problem?
It ignores such insignificant forces as time, entropy, and death
Acrofales
Profile Joined August 2010
Spain18291 Posts
Last Edited: 2016-02-18 19:20:02
February 18 2016 19:17 GMT
#59149
On February 19 2016 03:56 Plansix wrote:
Show nested quote +
On February 19 2016 03:48 Acrofales wrote:
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.

From a legal standpoint, that makes it very hard to prove a lot of things. If they can never open an Iphone for review, you could harass someone from it forever, sending death threats and other terrible thing and never be charged. As long as you did it all on the phone, including the creation of the address, it could be impossible to prove you sent the threats.

A tiny computer that can never be opened by anyone but the owner is the dream of almost every slightly smart criminal.


You're once again completely missing the point.

The police can:
1. Listen to what you're saying, and writing, through a wiretap they got a court order for. If those messages are encrypted, then whoever is on the receiving end is either receiving nonsensical jibberish, or can provide you with the decryption key.

2. The police can take that phone and figure out any number of ways to work around the security, just as today they manage to get info out of suspects in completely legal interrogations.

Neither of these have anything to do with the case at hand.

And the ability to store data in a way that nobody except the owner can access already exists and is freely available with a very brief Google search. All Apple did is take this already existing technology and embed it in their phone.

In actual fact, it already existed since virtually forever: you could always take your sensitive documents and hide them somewhere.

Edit: or as someone just pointed out: by memorizing it.
Plansix
Profile Blog Joined April 2011
United States60190 Posts
February 18 2016 19:20 GMT
#59150
On February 19 2016 04:11 WolfintheSheep wrote:
Show nested quote +
On February 19 2016 03:56 Plansix wrote:
On February 19 2016 03:48 Acrofales wrote:
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.

From a legal standpoint, that makes it very hard to prove a lot of things. If they can never open an Iphone for review, you could harass someone from it forever, sending death threats and other terrible thing and never be charged. As long as you did it all on the phone, including the creation of the address, it could be impossible to prove you sent the threats.

A tiny computer that can never be opened by anyone but the owner is the dream of almost every slightly smart criminal.

This exists already, and it's called the human brain.

Realistically, everything you send out on a phone still exists on the receiver's phone, and the records of some message being sent to that exact destination which align perfectly with the time stamps still exist. You don't need to break open a criminal's phone to prove that they were transmitting the incriminating messages beyond a reasonable doubt.

Honestly, people pretend that secrets didn't exist before computers, and that the situation is somehow worse. In reality it's far, far better for law enforcement, even with encryption, because you can't brute force or hack a person (legally) who is keeping secrets in their head. but a phone has no rights.


Courts require all evidence be authenticated by a witness, confirming that the evidence is what the attorney claims it is. So in the case of an email, they would need a witness that could testify that the email was sent from that specific phone while the witness is under oath. And that witness is open to cross examination, where the other side can as if they are 100% sure. If they can never open the phone, I question how they would ever prove that it came from that specific phone and not someplace else. The same goes for twitter and all other forms of social media that allow for anonymous log in.
I have the Honor to be your Obedient Servant, P.6
TL+ Member
WolfintheSheep
Profile Joined June 2011
Canada14127 Posts
February 18 2016 19:20 GMT
#59151
On February 19 2016 04:13 oneofthem wrote:
Show nested quote +
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

The problem, as stated a hundred times already, is that your use of the word "authorized" is not in line with reality.

You think it means "only the people we want".

The reality is "only people who have the key".

And there is a gigantic difference.
Average means I'm better than half of you.
Plansix
Profile Blog Joined April 2011
United States60190 Posts
February 18 2016 19:22 GMT
#59152
I was under the impression that if the FBI tries to brute force the phone or break the encryption, it will delete itself and destroy the evidence. Is that not true?
I have the Honor to be your Obedient Servant, P.6
TL+ Member
Acrofales
Profile Joined August 2010
Spain18291 Posts
Last Edited: 2016-02-18 19:29:08
February 18 2016 19:24 GMT
#59153
On February 19 2016 04:22 Plansix wrote:
I was under the impression that if the FBI tries to brute force the phone or break the encryption, it will delete itself and destroy the evidence. Is that not true?

It might be, but that has a VERY simple workaround. The real problem is the time it will take to crack the pass code.

EDIT: the work around is obviously to bypass the OS entirely and read out the ROM directly. You can keep flashing it ad infinitum if it keeps getting destroyed. I don't really see any way for iOS to get around that. Even if there's two separate ROMS like in their newer Secure Enclave systems. Only way I can think of doing this irreversibly is by tying the code up in some hardware and actually destroying that hardware if the code is typed wrong too often. Fairly certain that isn't happening.
oneofthem
Profile Blog Joined November 2005
Cayman Islands24199 Posts
February 18 2016 19:24 GMT
#59154
On February 19 2016 04:17 Gorsameth wrote:
Show nested quote +
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

How often do we need to lay out the path this takes before you understand why it is a problem?

it's not a problem with some relaxed assumptions. again a lot of intermediate and less scary optins that properly take into account compelling security/law enforcement interests
We have fed the heart on fantasies, the heart's grown brutal from the fare, more substance in our enmities than in our love
oneofthem
Profile Blog Joined November 2005
Cayman Islands24199 Posts
February 18 2016 19:25 GMT
#59155
On February 19 2016 04:20 WolfintheSheep wrote:
Show nested quote +
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

The problem, as stated a hundred times already, is that your use of the word "authorized" is not in line with reality.

You think it means "only the people we want".

The reality is "only people who have the key".

And there is a gigantic difference.

that is not what i think at all, try again. a front door works the same way but with proper safeguards that a 'back door' lacks. you might as well say the fact the owner can use the phone is a security flaw.
We have fed the heart on fantasies, the heart's grown brutal from the fare, more substance in our enmities than in our love
WolfintheSheep
Profile Joined June 2011
Canada14127 Posts
February 18 2016 19:27 GMT
#59156
On February 19 2016 04:20 Plansix wrote:
Show nested quote +
On February 19 2016 04:11 WolfintheSheep wrote:
On February 19 2016 03:56 Plansix wrote:
On February 19 2016 03:48 Acrofales wrote:
On February 19 2016 03:43 Jayme wrote:
On February 19 2016 03:35 Acrofales wrote:
Extremely relevant article from wired in 2014 on this issue. I agree completely. The sense of entitlement of law officials to be able to ruffle through your stuff is quite ludicrous.

Now I sympathize with the FBI in this case and wish them all the luck in the world cracking that phone. But just as we don't forbid ppl from using safes, or locking their front door, should we try to block the computational equivalent of these technologies.

And no, a search warrant doesn't mean you have to open your safe. Just your front door.


Edit: forgot the link.
http://www.wired.com/2014/10/golden-key/


That actually depends on what the Search Warrant says. If it specifically indicates a safe they know you own you're actually compelled to open it. Normally they just take the entire thing and open it themselves. Careful with statements like that. Search Warrants aren't limited to your house.



True, but they can't prove you didn't forget the code. And in this case the owner is dead, so can't be compelled to giving them the code.

From a legal standpoint, that makes it very hard to prove a lot of things. If they can never open an Iphone for review, you could harass someone from it forever, sending death threats and other terrible thing and never be charged. As long as you did it all on the phone, including the creation of the address, it could be impossible to prove you sent the threats.

A tiny computer that can never be opened by anyone but the owner is the dream of almost every slightly smart criminal.

This exists already, and it's called the human brain.

Realistically, everything you send out on a phone still exists on the receiver's phone, and the records of some message being sent to that exact destination which align perfectly with the time stamps still exist. You don't need to break open a criminal's phone to prove that they were transmitting the incriminating messages beyond a reasonable doubt.

Honestly, people pretend that secrets didn't exist before computers, and that the situation is somehow worse. In reality it's far, far better for law enforcement, even with encryption, because you can't brute force or hack a person (legally) who is keeping secrets in their head. but a phone has no rights.


Courts require all evidence be authenticated by a witness, confirming that the evidence is what the attorney claims it is. So in the case of an email, they would need a witness that could testify that the email was sent from that specific phone while the witness is under oath. And that witness is open to cross examination, where the other side can as if they are 100% sure. If they can never open the phone, I question how they would ever prove that it came from that specific phone and not someplace else. The same goes for twitter and all other forms of social media that allow for anonymous log in.

And if I wore gloves and sent these evil messages from newspaper cutouts like the good old days, you're saying I would never be convicted?

Sure, encryption is a road block for law enforcement, but again, the fact that the evidence even exists on a phone or device now instead of nowhere is a step up from even 20 years back, so let's not pretend that encryption is providing some magical hole that criminals can escape to.
Average means I'm better than half of you.
oneofthem
Profile Blog Joined November 2005
Cayman Islands24199 Posts
February 18 2016 19:30 GMT
#59157
On February 19 2016 04:15 puerk wrote:
Show nested quote +
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

creating it after the fact only works through a vulnerability in the system, so doing it shows the existence of said vulnerability: in this case there seems to be a way to force a locked iphone to update its firmware to a faulty one, that will give access

if this is possible is still unclear, but if apple does it, the systems flaw is exposed and others will copy the technique

the discussion i am interested in having concerns the general design of secured devices. this particular phone can just be individually unlocked, then apple can issue an update that would introduce front door measure at a later date for future problems.

i see no necessitation of security flaw from this sort of a design.
We have fed the heart on fantasies, the heart's grown brutal from the fare, more substance in our enmities than in our love
WolfintheSheep
Profile Joined June 2011
Canada14127 Posts
February 18 2016 19:32 GMT
#59158
On February 19 2016 04:25 oneofthem wrote:
Show nested quote +
On February 19 2016 04:20 WolfintheSheep wrote:
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

The problem, as stated a hundred times already, is that your use of the word "authorized" is not in line with reality.

You think it means "only the people we want".

The reality is "only people who have the key".

And there is a gigantic difference.

that is not what i think at all, try again. a front door works the same way but with proper safeguards that a 'back door' lacks. you might as well say the fact the owner can use the phone is a security flaw.

A secondary access point does not become magically more secure just because you called it a front door.

The security flaw is that there is an access vector that is identical across all devices outside of the control of the owner of the secured information.

Not that it has the word "back" in it.
Average means I'm better than half of you.
puerk
Profile Joined February 2015
Germany855 Posts
February 18 2016 19:32 GMT
#59159
On February 19 2016 04:30 oneofthem wrote:
Show nested quote +
On February 19 2016 04:15 puerk wrote:
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

creating it after the fact only works through a vulnerability in the system, so doing it shows the existence of said vulnerability: in this case there seems to be a way to force a locked iphone to update its firmware to a faulty one, that will give access

if this is possible is still unclear, but if apple does it, the systems flaw is exposed and others will copy the technique

the discussion i am interested in having concerns the general design of secured devices. this particular phone can just be individually unlocked, then apple can issue an update that would introduce front door measure at a later date for future problems.

i see no necessitation of security flaw from this sort of a design.


you expertly elaborate about issues you have not the first clue about. this might be the problem in this conversation: you have no fucking clue and act like you do and throw a tantrum when your dreamt up cluelessness solution doesnt work.

you behave like a child that talks about dry rain, and how parents are mean when they say that is not how rain works....
Acrofales
Profile Joined August 2010
Spain18291 Posts
Last Edited: 2016-02-18 19:37:57
February 18 2016 19:33 GMT
#59160
On February 19 2016 04:25 oneofthem wrote:
Show nested quote +
On February 19 2016 04:20 WolfintheSheep wrote:
On February 19 2016 04:13 oneofthem wrote:
On February 19 2016 04:08 puerk wrote:
On February 19 2016 03:56 oneofthem wrote:
On February 19 2016 02:35 Gorsameth wrote:
On February 19 2016 02:27 oneofthem wrote:
privacy hawks are making this harder than necessary. there are ways to make iphones unlock without thereby propagating said tools or methods. the means could itself be protected by encryption in some secure location.

strawmanning open access when it need not go so far is just a waste of time

Any backdoor in the OS itself is exploitable, its pretty impossible for it not to be.
Apple can do this with a 1 time program to only break this one phone.

Then the FBI/NSA/CIA will have another phone, and another, and another and....
Before long Apple has several people doing nothing but writing 1 use programs to break phones.

"Hey wouldn't it be easy if there was just a general backdoor we could use? We would need the physical phone so its all safe"

"Hey yeah that backdoor? Its kinda bad if we need the phone itself you know, we want to catch people before they commit crimes so we need to be able to remote breakin"

Oh hey look someone broke the Iphone backdoor and now everyone's data is on the streets, evil Apple let this happen!.

Intelligence organizations have a long history of breaking laws and gathering anything and everything they can get their hands on regardless of use. No I don't trust them to act responsibly with this technology for even a second.

the whole point of bringing it up in a court of law is to have a front door rather than a backdoor, with the proper precautions and safeguards.

there is already a front door, the passkey the owner of the phone used. (have they even tried swiping his finger on the sensor, if it is biometrically locked?)
everything else is by definition of encryption systems a back door
so what is the problem with creating another similarly secure front door mechanism that authorized access can open?

The problem, as stated a hundred times already, is that your use of the word "authorized" is not in line with reality.

You think it means "only the people we want".

The reality is "only people who have the key".

And there is a gigantic difference.

that is not what i think at all, try again. a front door works the same way but with proper safeguards that a 'back door' lacks. you might as well say the fact the owner can use the phone is a security flaw.


OK. So the front door requires a special key that the user can configure. He sets his code to fjord$^_fhrid4568nrtbr÷==AT&TIEN

He memorized this key and doesn't tell anybody. He subsequently dies, leaving the front door locked forever.

Now what you're proposing is that somehow Apple magically has this key too? How?

EDIT: and just to be clear this, and only this is the front door. Amy other access mechanism is, by its very definition, a back door.
Prev 1 2956 2957 2958 2959 2960 10093 Next
Please log in or register to reply.
Live Events Refresh
GSL
08:00
2026 Season 1: Playoffs
Cure vs herOLIVE!
SHIN vs Maru
GSL EN (SOOP)0
LiquipediaDiscussion
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
Rex 29
StarCraft: Brood War
Mind 432
Killer 379
scan(afreeca) 114
Zeus 68
Hm[arnc] 58
Larva 50
Backho 47
Sacsri 31
JulyZerg 26
Bale 25
[ Show more ]
PianO 19
Noble 14
Sharp 6
Dota 2
XaKoH 414
NeuroSwarm164
League of Legends
JimRising 613
Other Games
summit1g11317
WinterStarcraft517
monkeys_forever281
Organizations
Counter-Strike
PGL16476
Other Games
gamesdonequick778
StarCraft 2
IntoTheiNu 274
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
[ Show 16 non-featured ]
StarCraft 2
• Berry_CruncH201
• CranKy Ducklings SOOP47
• LUISG 22
• AfreecaTV YouTube
• intothetv
• Kozan
• IndyKCrew
• LaughNgamezSOOP
• Migwel
• sooper7s
StarCraft: Brood War
• BSLYoutube
• STPLYoutube
• ZZZeroYoutube
League of Legends
• Lourlo1026
• Jankos871
Counter-Strike
• C_a_k_e 1667
Upcoming Events
IPSL
7h 36m
Bonyth vs Napoleon
G5 vs JDConan
BSL
10h 36m
OyAji vs JDConan
DragOn vs TBD
Replay Cast
1d
Monday Night Weeklies
1d 7h
Replay Cast
1d 15h
The PondCast
2 days
Kung Fu Cup
2 days
GSL
3 days
Replay Cast
3 days
GSL
4 days
[ Show More ]
WardiTV Spring Champion…
4 days
Replay Cast
4 days
Sparkling Tuna Cup
5 days
WardiTV Spring Champion…
5 days
Replay Cast
5 days
RSL Revival
6 days
Classic vs SHIN
Rogue vs Bunny
BSL
6 days
Replay Cast
6 days
Afreeca Starleague
6 days
Liquipedia Results

Completed

Escore Tournament S2: W7
WardiTV TLMC #16
Nations Cup 2026

Ongoing

BSL Season 22
ASL Season 21
IPSL Spring 2026
KCM Race Survival 2026 Season 2
Acropolis #4
KK 2v2 League Season 1
BSL 22 Non-Korean Championship
SCTL 2026 Spring
RSL Revival: Season 5
2026 GSL S1
Heroes Pulsing #1
Asian Champions League 2026
IEM Atlanta 2026
PGL Astana 2026
BLAST Rivals Spring 2026
IEM Rio 2026
PGL Bucharest 2026
Stake Ranked Episode 1
BLAST Open Spring 2026
ESL Pro League S23 Finals
ESL Pro League S23 Stage 1&2

Upcoming

YSL S3
Escore Tournament S2: W8
CSLAN 4
Kung Fu Cup 2026 Grand Finals
HSC XXIX
uThermal 2v2 2026 Main Event
Maestros of the Game 2
WardiTV Spring 2026
2026 GSL S2
BLAST Bounty Summer 2026
BLAST Bounty Summer Qual
Stake Ranked Episode 3
XSE Pro League 2026
IEM Cologne Major 2026
Stake Ranked Episode 2
CS Asia Championships 2026
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2026 TLnet. All Rights Reserved.