tag:blogger.com,1999:blog-173291792024-03-08T12:44:53.941-08:00Email-Security<b>Welcome to Email-Security.</b> This is a technical development forum dedicated to a fresh exploration of the Internet email security issues of today, with website at <a href="http://email-security.net">http://email-security.net</a>. Comments and paper contributions on the theme of email security are welcome. Papers will be peer-reviewed before publication. Product and service listings are also welcome, see website.Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.comBlogger52125tag:blogger.com,1999:blog-17329179.post-3498330674288299482011-07-17T23:00:00.000-07:002011-07-17T23:03:01.035-07:00Electronic Medical Records -- the new email frontier?The statistics regarding the impact of EMR (Electronic Medical Records) on efficiency seem far less than encouraging. As commented privately to me, in more than one occasion, significant net drops in provider efficiency are the norm, despite the hype. <br />
<br />
The push for EMR adoption on the part of the government and payers is about data, not provider productivity (at least not the kind of productivity that improves a provider's bottom line). A 6% penalty on Medicare billing or a $64,000 incentive is small compared to the real cost of EMR adoption for many providers. The gap between what providers are being sold on and reality could hardly be broader in the industry...<br />
<br />
So, the reality is that today's EMR is a really hard thing to do well, as also evidenced by the reported high numbers of failed EMR conversions and large number of second EMRs (replacing a first EMR).<br />
<br />
<b>Electronic Medical Records -- the new email frontier?</b><br />
<br />
In contrast, using encrypted email can be much easier for adoption and even provide an EMR basis <b>IF</b> it can complement rather than replace an email system. It should require no changes in the user GUI, no learning, no commitment in other areas, no change to any other aspect of the operation, and should have low cost compared to the base system (email).<br />
<br />
This is exactly what ZSentry at <a href="http://zsentry.com">http://zsentry.com</a> has been providing, as a complement system to Google Apps, Outlook, iPad, and other <i>existing</i> email solutions, enabling HIPAA and HITECH Safe Harbor (the next need after HIPAA) practically "out-of-the-box".<br />
<br />
And ZSentry can also be used as part of a local EMR, allowing secure communication in and out of it. One such hybrid ZSentry/EMR system has recently been certified by the DHHS in the US ARRA program.<br />
<br />
So, Yes, albeit the statistics on current EMR efficiency are not encouraging, we can start with simpler systems that are, well, simpler to work with.<br />
<br />
<b>How about the future?</b><br />
<br />
If we see the statistics of CNC (Computer Numerical Control) adoption in the early 50's in the US, in the 70's for word processors, and other IT application examples, including the Internet itself, we will see the same trend of less efficiency, frustration, followed by exponential, unbelievable at the time, improvements in efficiency, cost, quality, and every other parameter (except when people look at direct employment reduction as a downside, rather than as an opportunity for progress).<br />
<br />
Now, ask yourself how much would it cost to run an IT company the way a physician runs his office today?<br />
<br />
For example, how many people-hours/day are used faxing in a practice today? And faxes are not even HIPAA-compliant...How long can this waste go on?<br />
<br />
It's just natural that <i>the future comes with a shock-wave.</i> But automation-driven efficiency improvements grow fast and are worth the trouble. The U.S. Census Bureau took eight years to complete the 1880 census, but just one year for the 1890 census. The difference then: automatic tabulation with Hollerith cards.<br />
<br />
Now, fast-forward to the US health-care sector in 2011. What's happening with the SOHO and SMB health-care markets? They are shrinking fast since ca. 2000, and becoming far more competitive in price. For example, the share of solo practices among members of the American Academy of Family Physicians fell to 18 percent by 2008 from 44 percent in 1986. And US census figures show that in 2007, just 28 percent of doctors described themselves as self-employed, compared with 58 percent in 1970. Many of the provisions of the new health care law are likely to accelerate these trends in the health care sector.<br />
<br />
Because efficiency will become even more critical for the survival of solo and small practices, they may be more likely to use simpler, secure email first and move to EMR as possible. I doubt they will continue to use fax, or even dare to ask their patients to run to the pharmacy with a written prescription, or refill by phone. Larger health-care providers may also see faster returns by using simpler systems first. Keep It Simple -- and I see no need for adding a second S at the end. <br />
<br />
Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com2tag:blogger.com,1999:blog-17329179.post-76555548130199000022011-06-29T15:45:00.000-07:002011-07-02T17:41:25.416-07:00Email vs Postal Mail Security<i>"If can send this document by postal mail, why can't I email it?"</i><br />
<br />
Saying that email is like a postcard, that anyone can read and even write on, is usually not enough to highlight the differences. After all, there are so many emails out there, people think, that email is "protected by numbers." And anyone can also open my envelope and read my postal mail. <br />
<br />
<b>What's the difference?</b><br />
<br />
To further clarify the difference, it is useful to consider the different roles of an email service provider versus a postal mail or courier service in terms of security of medical records. A health care organization (HCO) must comply with HIPAA (medical records privacy rules) to assure the privacy of a patient's Protected Health Information (PHI). To send PHI, the HCO has a legal obligation to sign a Business Associate Agreement (BAA) with any service provider that may become aware of the PHI, assuring that the PHI will be protected as determined by HIPAA.<br />
<br />
The US Post Office and courier services do not need to have a BAA to transfer mail. It is legally understood in HIPAA terms that the usual "sealed envelope" is enough to protect the PHI and deny potential access to it by the service provider -- even though access is clearly possible by the service provider personnel, and would require just opening the envelope. <br />
<br />
However, "opening the envelope" would have to be done by someone first going through the physical access controls of the mail provider, is a physical act that can be done only one envelope at a time, takes time and space, can be observed and reported by a co-worker or someone else (including video surveillance). Typically, several people need to be involved and would need to collude in an entire mesh of secret events. Further, "opening the envelope" is still likely to leave traces that the recipient can see.<br />
<br />
This is not the case with electronic records, such as electronic PHI (ePHI). A factor that militates against ePHI is the fact that although paper and electronic records are both vulnerable to subversion, it is a lot easier and faster to read what is in an electronic record somewhere, or search for specific information, than it is to read or search what is on paper in a closed envelope.<br />
<br />
HIPAA enforcement and patients are also well aware that electronically one can read, search, and even modify a million records with as little as a few keystrokes, for example from a cybercafe in Moldavia or somewhere in China. <br />
<br />
These are critical vulnerabilities that need to be addressed when using ePHI - for example, that mass disclosure and potential subversion can be so readily accomplished from the safety of a remote laptop or phone, and that it would be unavoidable or even undetectable.<br />
<br />
To answer these concerns, the email provider must demonstrate a number of barriers, including defense in depth. This should be based not only on cryptographic assurances, but how they are distributed, and how ePHI entering the email provider is bound to communication channels in a manner that the ePHI is rendered demonstrably inaccessible to an attacker, both through physical access controls and through cryptographic protocols. <br />
<br />
Moreover, the email provider should include a step-by-step description of the process, available for auditing, so that when someone asks, "What if the intruder succeeds in breaking into the system to change X?" this can be clearly answered, for example, by (i) to change X would cause a subsequent binding failure, thus it would be detectable except with parallel access to Y and Z, which are independently inaccessible; or (ii) knowledge of an alternate (and attacker-desirable) value for X is computationally impossible to achieve during the PHI lifetime period, and the effort could not be leveraged to any other X.<br />
<br />
In short, the email provider should use not just cryptography but also other techniques to make it is as impossible as desired to tamper with the PHI. This includes server actions to protect each client and to protect the servers from malicious use of the clients.<br />
<br />
In addition to the foregoing barriers, "opening the envelope" by the email service provider staff or online attackers would require having the keys to do so. However, where and how to securely keep the keys from insiders, and who watches the watchers?<br />
<br />
Looking at NASDAQ, Epsilon, Citibank, and other high-profile "secure service provider" breach cases in 2011, shows a bleak view as not even large organizations could prevent such attacks.<br />
<br />
However, these requirements can be easily met by an email service provider using the ZSentry Sans-Target technology, because online servers and personnel simply do not have the keys to do it, there is no copy of the keys anywhere, and the keys are too large for a trial-and-error attack to succeed in practice. <br />
<br />
Further, ZSentry treats the email body as a "data blob" and does not scan it. There is also no condition whereby the ZSentry service or personnel could be made aware of the ePHI, providing even stronger protection for an email service provider using ZSentry services for encryption.<br />
<br />
Therefore, in addition to assuring that the ePHI is encrypted and de-identified, an email service provider can use the ZSentry technology to technically prevent anyone, including own personnel and online attackers, from breaking the encryption or the de-identification.<br />
<br />
More info in answer to the question "Why is ZSentry secure?" in <a href="http://zsentry.com/security-email-faq.htm">http://zsentry.com/security-email-faq.htm</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-44508289685598282462011-05-28T09:31:00.000-07:002011-06-08T09:56:39.074-07:00Silverlight, Flash, PDF and other Online Form Techniques<i>[Target Audience: Developers and System Administrators]</i><br />
<br />
There seems to be growing interest for an online form service allowing customers to collect input from users and send it back as structured information to organizations, while preserving confidentiality including input, transmission, archive, access, and auditing functions.<br />
<br />
The ability to enable customers to style online forms, for different server scenarios including absence of server-SSL input protection, should also be part of the available customer-design process. Further, the online form service should allow customers to use web forms, email forms, SMS forms, storage file forms, and any other communication process.<br />
<br />
ZSentry is currently beta-testing such a <i>Secure Online Form</i> service that meets these considerations.<br />
<br />
This work has led us to think more deeply about the pros and cons of Silverlight, Flash, PDF and other methods used to provide online form functions, as well as security and usability, which are the motivations for this posting and feedback.<br />
<br />
In many cases of interest, the online form service must assure full HIPAA compliance. Safe Harbor compliance (even without HIPAA) is also coming to the foreground given the high profile and costs of recent breaches such as with Epsilon and Sony.<br />
<br />
From the security viewpoint, and contrary to ZSentry for Google Apps (where both the need for HIPAA and Safe Harbor compliance can be satisfied by ZSentry), online forms as enabled by Google Docs cannot at this time be used by ZSentry as a basis for such secure service. The recent security breaches caused by phishing attacks against Google Apps and Gmail further highlight the concerns of storing plaintext information in online services such as Google Docs.<br />
<br />
Our usability starting point is that online form functions are "critical" and, thus, in principle should be usable without asking the user to download or do anything else. <br />
<br />
Usability is also necessary to ensure security, as we mention elsewhere (Usability is a system's #1 Security Property). In short, a secure online form service that is secure but not usable will not be used and, therefore, will actually fail to be secure. If security is too difficult or annoying, users may give up on it altogether.<br />
<br />
Therefore, while it is OK for an information or help document to be available in PDF (with all the PDF requirements/problems to be met and solved at the user side), users can either view the PDF using any quick view software/service (including Google) or just move on to something else. Not so with an online form that needs to be entered.<br />
<br />
We also see that many users are just that -- users -- and have no admin privilege in their devices. Increasingly, devices are locked (also by manufacturers -- see iPad vs Flash) for installation.<br />
<br />
These concerns go against Silverlight. It could be better to use Flash (more widely available) but still many users would be left out as new Flash versions come out quite frequently and many users cannot update or do not like to bothered by yet another update. Silverlight also suffers from frequent updates. <br />
<br />
How about JavaScript? It is commonly available. However, even JavaScript should only be used to provide added usability, while the online form itself should be functional without JavaScript. This is important not just because some devices have problems with JavaScript; some networks and services also block JavaScript.<br />
<br />
In comparison, even though Silverlight or Flash based online forms may look more usable for corporate intranets and other special subsets of users, for public Internet users they are actually not as usable as HTML/CSS with optional JavaScript.<br />
<br />
ADA and accessibility requirements also point to HTML/CSS with optional JavaScript.<br />
<br />
<b>In conclusion:</b><ol><li>Knowing the target-user comes first, as it defines how the message needs to be delivered. Silverlight, Flash, PDF forms, and other non-HTML solutions can be used well in an intranet (internal services in a hospital, for example) where the organization can control the server, the network, the filters, all the clients, and what is installed and updated at any moment in those clients. Lacking that inclusive control, it is bound to fail in the public Internet for the same reasons.</li>
<li>From the user's viewpoint, HTML/CSS with optional JavaScript should be the best practice for online forms using the public Internet, as providing the needed functions while not standing in the way of using the online form -- which is the user's goal.</li><li>The ZSentry API for secure online forms is designed so that it can be used with any desired method, including HTML/CSS with optional JavaScript, Silverlight, Flash, PDF forms, and other interfaces. It is up to the developer to make the best use of it.</li>
</ol><br />
Comments are welcome. ZSentry for secure online forms is currently in beta test. Please submit a <a href="https://zsentry.com/support-ticket.htm">Support Ticket</a> if you are interested in participating now or when the service is launched.<br />
<br />
Cheers,<br />
Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com2tag:blogger.com,1999:blog-17329179.post-47833244930201036692011-04-23T09:31:00.000-07:002011-06-07T12:15:48.816-07:00NMA ZSentry API -- Frequent Questions<i>[Target Audience: Developers and System Administrators]</i><br />
<br />
The NMA ZSentry API is a secure software service that works in your environment. You can use the API both in a server as well as in a client context. There is no plugin or installation. With the ZSentry API you can readily connect, send, and store information securely, for example as a secure email, enabling secure communications in desktop, cloud, and mobile platforms, anywhere.<br />
<br />
<b>What is the NMA ZSentry API?</b><br />
An SMTPS (SMTP over SSL, also called implicit SSL) service. You use the ZSentry API to connect your system to the ZSentry engine using TCP port 465 for encryption and decryption directly or through a message service or application. The resulting secure emails are encrypted per message and can be sent using ZSentry's own SMTP server, without requiring SSL, or can use another resource as a mail relay with or without SSL.<br />
<br />
<b>Why port 465?</b><br />
With port 465 there is no ZSentry data exchange without SSL, including the ZSentry Usercode and Password data. Contrary to the submission port 587 with STARTTLS where the connection changes from non-SSL into SSL, port 465 has SSL-at-connect required. We prefer port 465 because every connection is encrypted from the start, preventing problems such as the SSL bridge attack (a well-known problem also for web browsers).<br />
<br />
<b>What server-side applications and servers are compatible?</b><br />
The ZSentry API complies with IETF standards. ZSentry tested the API using servers with the dovecot/postfix Debian implementation and other Linux flavors. Customers have tested the API with a variety of Linux and Windows technology, including System.Net.Mail, CDOSYS (Collaborative Data Objects, Cdosys.dll), VB6, PHP, Java, and Sybase PowerBuilder.<br />
<br />
<b>Do users need to use ZSentry App (the web interface)?</b><br />
No, not necessarily. The ZSentry API sends secure email that can be read and decrypted using the ZSentry App as well as leading Mail clients without any plug-in. This means that with the ZSentry API there is no web browser security and usability limitation for users, and users do not need to use the ZSentry App for reading, sending and saving secure messages. However, the ZSentry App can be used in a webmail environment.<br />
<br />
<b>How about Google Apps and Gmail?</b><br />
Web mail services including Google Apps and Gmail can be set up to send HIPAA-compliant secure email using the ZSentry API directly, as a cloud Mail client with Single-Sign-On. This can be done by users, without an add-on or installation. Reading and saving secure email can also be done through Google Apps and Gmail,<br />
using ZSentry App in the webmail environment or a Mail client.<br />
<br />
<b>How are Mail clients used?</b><br />
Mail clients can be set up to send, read, and save HIPAA-compliant secure email using the ZSentry API directly, with Single-Sign-On. The user only sees the Mail client GUI for sending, reading and saving messages. This can be setup by users, without a plug-in or installation. A web browser is still used on the client side, but it is used transparently as a middleware and not as a GUI.<br />
<br />
<b>What clients are compatible?</b><br />
The ZSentry API complies with IETF standards. Compatible clients include any major web browser, Outlook, Outlook Express, Windows Mail, Apple Mail, Entourage, Thunderbird, iPhone, Android, and iPad, on Windows XP, Vista, Windows 7, Linux, Mac OS X and iOS.<br />
<br />
<b>What applications can be used at the client side?</b><br />
Clients can use several applications including webmail, email, IM, SMS, and file storage services.<br />
<br />
<b>Can I debug the SSL/SMTP flow?</b><br />
Yes. This is a typical SSL/SMTP trace enabled by the API and which you can also capture in normal use to log delivery evidence for your service:<br />
--------------<br />
SMTP -> FROM SERVER:220 mail.example.com ESMTP Postfix<br />
SMTP -> FROM SERVER: 250-mail.example.com 250-PIPELINING 250-SIZE 20480000 250-VRFY 250-ETRN 250-<br />
AUTH PLAIN LOGIN 250-AUTH=PLAIN LOGIN 250-ENHANCEDSTATUSCODES 250-8BITMIME 250 DSN<br />
SMTP -> FROM SERVER:250 2.1.0 Ok<br />
SMTP -> FROM SERVER:250 2.1.5 Ok<br />
SMTP -> FROM SERVER:354 End data with .<br />
SMTP -> FROM SERVER:250 2.0.0 Ok: queued as E519A5801F4<br />
-------------------------<br />
<br />
More Information:<br />
<a href="http://zsentry.com/zapi.htm">http://zsentry.com/zapi.htm</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-75313172401851296872011-03-22T16:54:00.000-07:002011-04-22T19:35:08.894-07:00Sender-Defined SecurityWhen you send an email to a recipient, who bears the highest risk? <br />
<br />
Usually, you. Because the information is transferred from sender to recipient, the sender likely has the highest risk. The information may be valuable to both, but the sender is the party that can be found liable for (even if inadvertently) disclosing it to others. Email is risk-asymmetric. <br />
<br />
For example, patients may initiate communications with a health care provider using unencrypted email, according to HIPAA. This is done at the patient's risk. However, providers may have concerns when replying to the patient using unencrypted email. Providers are subject to potential liability under the HIPAA Privacy Rule [see 45 C.F.R. § 164.530(c)], including their determination whether a patient is aware of the possible risks of using unencrypted email.<br />
<br />
Since the sender is the party at risk, the sender needs to be able to define the security conditions, for example, for confidentiality, delivery, and tracking of the message.<br />
<br />
Let's see how certified postal mail works. The sender chooses the service and the type of envelope to protect the message. The sender also chooses the instructions<br />
that must be followed before the envelope can be opened by the recipient. The recipient has no charge to pay for, or burden, in order to receive the envelope, and does not have to do anything before the envelope is sent. The recipient is able to verify the identity of the sender and, if so desired, refuse the envelope. The recipient can open the envelope if and only if the recipient is willing to follow the sender's instructions (e.g., providing name, address, date, signature).<br />
<br />
The same should be available for sender-defined security modes of operation for email. <br />
<br />
Using IBE encryption (Identity-Based Encryption, Voltage, MessageGuard) does not work because IBE requires sender key-escrow, so that the sender is always at risk that her messages can be read by someone else.<br />
<br />
Clearly, a conventional public-key model (PKI, X.509, PGP) cannot provide this functionality either. The security model for protecting email with public-key cryptography is backwards, technically and business wise. The sender, who is the party at risk, has to trust a lock provided by the recipient (his public-key) to protect her secrets (the message). <br />
<br />
If certified postal mail would provide message security a la PGP, PGP/MIME or S/MIME email, the sender would have to convince the recipient to pay and send in advance an envelope for the sender to use. The sender, however, would never know whether the envelope indeed prevented others from prying into its contents. Moreover, the recipient is not the business driver who needs to provide, pay for and protect the lock. The sender is the party who has the motivation to spend money to provide and protect the lock.<br />
<br />
These are well-known and recognized standards for encryption of email. However, their security and usability problems noted above cannot be solved by improvements in usability, better pricing or user education regarding how IBE, PKI, X.509, or PGP work. It simply does not work as it should work. That's also why conventional secure email has been difficult to use. It has the wrong logic for the problem.<br />
<br />
<a href="http://zsentry.com">ZSentry</a> is a security technology that was developed after these standards and improves upon them in both usability and security. ZSentry provides per-message encryption, authentication, message control, and auditing, protecting information in motion and at rest. ZSentry reduces the trust and control requirements in several critical areas, making it easier to attain and demonstrate a higher level of security while increasing usability. <br />
<br />
Most importantly, ZSentry enables a secure email system that works closer to how postal mail, including courier and certified postal mail, works.<br />
<br />
Because ZSentry mail can be delivered securely to both registered and also to unregistered users, the sender is able to define the <a href="http://zsentry.com/email-security/help_em.htm#dp">delivery conditions</a> that must be satisfied by the recipient before the message can be decrypted (delivered) in each case. <br />
<br />
For example, based on what is at stake and what threats may apply, versus the usability requirements, the sender can select from the following delivery conditions:<br />
<br />
<b>Require Registration:</b> (For higher security: use this option) The recipient must register before reading the message. After the recipient registers, delivery is further controlled by the Delivery option specified by you for registered users (see next items).<br />
<br />
<b>Require Login:</b> (For higher security: use this option) The recipient must be registered and login before reading the message. To reduce risk, this process includes mailbox authentication, login monitoring, message expiration, and other control features.<br />
<br />
<b>Read Until Expiration:</b> (For higher usability: use this option) Available for both registered and unregistered users. The recipient is allowed to decrypt the Zmail, including attachments (if any), as many times as desired until the message expires. The security of the Read Until Expiration mode is based on mailbox authentication, login monitoring, and expiration control. The sender can choose when to expire, where the sooner the better for security.<br />
<br />
<b>Reply to Sender:</b> (available in all options, including without registration or login) This is a very useful choice to receive a secure reply, as the recipient has no burden to reply securely (can reply securely just with one-click).<br />
<br />
The security of these Delivery choices is further enhanced by <a href="http://zsentry.com/email-security/help_em.htm">other choices</a>, including for Control and Tracking, for example allowing the sender to choose a message self-destruct time and real-time audit reporting with a Return Receipt. <br />
The Return Receipt is sent back to you with information on When, Where, How, and by Whom your message was read. <br />
<br />
You can also compare the Return Receipt with the corresponding Send Receipt (available from the server, upon sending), closing the loop with potentially more comprehensive delivery and notary evidence than using courier or certified postal mail, with less cost and immediate delivery.<br />
<br />
In summary, <a href="http://zsentry.com">ZSentry</a> puts the sender in the driver's seat of secure email, with more and more effective options than any other system including IBE, PKI, X.509, PGP, courier and certified postal mail.Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-62364518939785983002011-02-26T15:33:00.000-08:002011-04-22T16:17:33.271-07:00How about password security?Password security may look like yet another oxymoron, as two intrinsically contradictory terms. If it is a password, how can it be secure?<br />
<br />
Indeed, there are many and valid security arguments not to use passwords. You can find a summary in the paper <a href="http://email-security.net/papers/takefivecompact.htm">Take Five In Internet Security</a> where, in a five-minute appraisal, we present and discuss a typical dialogue with someone claiming that they have a secure server while using username/password authentication.<br />
<br />
However, not all passwords are created equal. ZSentry, for example, uses two-factor, strong authentication without passwords -- even though, for familiarity, the second-factor is called ZSentry Password. <br />
<br />
This is important for usability, as passwords are easy to use. <br />
<br />
The security difference is that ZSentry does not store your Usercode or Password, not even encrypted. This technology, called ZSentry Sans-Target, solves the vulnerability problems of common passwords, including dictionary attacks and password files, as described in the study <a href="http://email-security.net/papers/takefivecompact.htm">Take Five In Internet Security</a>. <br />
<br />
The ZSentry technology eliminates common online targets such as username/password lists, names, email addresses, plain text user data, meta-data, and even the encryption/decryption keys themselves, while adding two-factor mutual authentication, adaptive security, and password-hardening.<br />
<br />
The critical point is that ZSentry Sans-Target provides the best defense possible against data theft -- the best defense against data theft is to not have the data in the first place. <br />
<br />
Another important point is that ZSentry is not as dependent on password quality for security, as conventional systems are. With ZSentry technology, passwords are paired with an unpredictable ZSentry Usercode, and both are not at risk anywhere. <br />
<br />
Nonetheless, it is recommended that ZSentry passwords should include at least one control or punctuation character. All of the characters !@#$%^&*()_-+=[]|\;:"?/,.< >`~' and space can be used in ZSentry passwords (space cannot be used at the beginning or end of a password).<br />
<br />
For advanced password use, for maximum protection, you can use ANSI codes from #32 to #255 (keyboard ALT-number, no space at the start or end of a password). This simple strategy enables more than 132 bits of entropy with just 13 ZSentry Password characters (and the Usercode). <br />
<br />
The conventional difficulty for using ANSI codes in passwords is solved by the ZSentry function Password Peek (during signup and login). You can easily see and verify what you typed before you submit, even for ANSI CODES such as ALT-0159 for Ÿ (Latin Capital Letter Y With Diaeresis).<br />
<br />
With ZSentry, you can now use passwords as securely as you want.Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-23763217370490019442011-01-30T16:28:00.000-08:002011-02-09T17:24:23.740-08:00The Technical Problem of Information Security<i>This article provides an in-depth look into Information Security and End-to-End Security. What is the problem, what can we do about it, and how effectively?</i><br />
<br />
This discussion presents the technical requirements for an End-to-End IT security solution, summarized in four points:<ol><li>integrate core security services;</li><li>eliminate known weak or costly links such as password lists, access control databases, shared secrets, and client-side PKI;</li><li>avoid the seemingly desirable scenario of a single point of control, which is recognized as a single point of failure;</li><li>bind a coherent system of trust to the solution and to its users.</li></ol>The NMA ZSentry solution is mapped as an example of a system that complies to these requirements and provides a cost-effective implementation. <br />
<br />
The technical problem of information security can be stated, quite simply, as <i>"to avoid too much concentration of information and power, while allowing enough information and power so as to make a task possible to execute." </i><br />
<br />
For example, an all-knowing, all-powerful entity would be the perfect attacker and could break any security measure.<br />
<br />
That is why we oftentimes talk about "need to know" and "separation of powers." We name these principles, respectively, information granularity and power granularity. <br />
<br />
Such principles mean that information should not be provided in its entirety to a single entity and one should also avoid the scenario where any entity is, at the same time, user, administrator and auditor. That is why business information and power should be carefully divided, for example, among local employees, the office management, the enterprise management and the customer.<br />
<br />
Also, contrary to what is oftentimes advocated in regard to IT security solutions, there should be no single point of control in an IT security system, which point must be recognized as a single point of failure -- i.e., <i>no matter how trustworthy that single point of control is, it may fail or be compromised and there is no recourse available because it is the single point of control.</i><br />
<br />
One of the earliest references to this principle can be found some five hundred years ago in the Hindu governments of the Mogul period, who are known to have used at least three parallel reporting channels to survey their provinces with some degree of reliability, notwithstanding the additional efforts.<br />
<br />
Tools such as authentication and authorization can help define information and power granularity. However, at its most basic level, a secure IT system needs to do much more than just control authentication and authorization.<br />
<br />
No matter how much assurance is provided that each component of a secure system is correct, when operational factors such as collusion, internal attacks, hackers, bugs, viruses, worms or errors are taken into account, the system may fail to be effective -- i.e., may fail to be secure in the context of its operational use. <br />
<br />
In addition, underlying assurance problems such as insecure operating systems and reoccurring buffer overflow vulnerabilities are not likely to improve over the coming years.<br />
<br />
There is a real need, thus, to bring together policy, management and implementation considerations that could influence effectiveness assurances for each particular IT security solution. <br />
<br />
Other security principles such as redundancy, diversity, and least privilege also need to be used in defining the specific requirements for a secure IT system. In addition to being specific, such requirements need to be clearly formulated, decidable and, as much as possible, complete. <br />
<br />
Additionally, an end-to-end design is important to assure effectiveness, because attacks and errors are hard to detect and prevent at the interface points. End-to-End (E2E) security is defined as safeguarding information in a secure telecommunication system by cryptographic or protected distribution system means, from point of origin to point of destination<br />
<br />
For lack of paper trails or other material proof, non-repudiation is also often essential for Internet and IT security systems. A common definition states that non-repudiation is about providing proof that a particular act was actually performed, for example as demonstrated by a trusted time-stamp. However, this definition is etymologically incorrect and, thus, should be repudiated. It also misses the point. We and other authors prefer to define non-repudiation as <i>"preventing the false denial of an act"</i>. In other words, it is not enough to provide proof of an act, as a denial may be presented and create a tie. While a correct denial should win the tie, a false denial should not. <br />
<br />
To be effective, non-repudiation needs to take into account technical and business considerations. For example, bank checks are non-repudiable in the sense that a check is paid if (1) you did not tell the bank beforehand that the check should not be paid and (2) the signature does not look like a signature you did not make. The reader should note the double-negative, which provides less room for customer repudiation -- the signature does not have to look like a signature that you made, it just has to not look like a forgery.<br />
<br />
Commonly, IT secure systems also must satisfy security standards such as the Controlled Access Protection (C2) level of security as defined in DoD 5200.28-STD (the Department of Defense Trusted Computer System Evaluation Criteria), the ITSEC Level 3 as defined in the Common Criteria for Information Technology Security Evaluation, or the Code of Practice for Information Security Management BS 7799 (a British Standard that is the basis of the ISO/IEC 17799-1 Standard).<br />
<br />
To meet these objectives, an effective IT security solution should be based on two points:<ul><li>Clear security principles, algorithms and products based on time-proven designs; and</li><li>Independent, permanent verification and validation of the systems’ security features.</li></ul><br />
The first point defines the component quality used in the IT system, where a weak component may compromise the whole system. For example, a system that relies on conventional username/passwords for authentication is likely already broken. <br />
<br />
The second point focuses on the need to continuously evaluate all potential and existing threats, and verify any additional security design features that might be necessary to mitigate the risks stemming from the most likely and/or most damaging threats associated with the customer environment, and eventual changes in that environment.<br />
<br />
The ZSentry security solution satisfies both points. ZSentry uses standard encryption technology with the unique <a href="http://zsentry.com/hsecurity.htm">Sans Target</a> method to keep it safe and HITECH Safe Harbor compliant to send data between parties as regular email. It does not require installation of any software, which is great for both security and usability, and it even adds functionality such as self-destruct, with message level access control. This is also good news for Novell users who have become used to this type of functionality. <br />
<br />
For example, using ZSentry renders your private health data unreadable to any third party who might come across it, and takes the issue of HIPAA breaches and fines off the table.<br />
<br />
Read more about <a href="http://email-security.net/papers/takefive.htm">Why Not Passwords?</a><br />
<br />
Read more about <a href="http://nma.com/papers/e2e-security.htm">End-to-End IT Security</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-57939564765386691212010-12-30T04:32:00.000-08:002011-01-24T14:06:52.138-08:00Prevent Attacks and Improve Usability By Understanding Asymmetric Threats<i>How asymmetric threats can blindside your organization, and how to prevent attacks supporting productivity.</i><br />
<br />
Users are typically blamed as the weak link in security. Yet, we also need to take into account that users are often just naively reacting to <i>asymmetric threats</i>, that are difficult to evaluate even by experts.<br />
<br />
<b>Asymmetric Threat:</b> a low-probability event with high-consequence. <br />
<br />
The difficulty with preventing this type of threat is that even though the consequences could be disastrous (including collateral damage, information leak, and reputation loss), users may mistakenly accept the risk solely based on the event's low-probability of failure. As known in the medical field for an elective procedure called VBAC, many patients choose to undergo the procedure because the risk is low (say 1%), even though the consequence can be death or serious life impairments for both mother and child.<br />
<br />
For example, in spite of all the attacks that we hear about on the Internet, it is known that the probability that any specific location will be successfully attacked is likely to be low for organization users. Consequently, users commonly see and rate online attacks as low-probability events ("Attackers will not guess my password.") and are not usually as concerned as they should be about the high-consequence side. <br />
<br />
As another example, consider phishing. It is a typical asymmetric threat, where the user is asked to do something that seems safe (low-probability of failure) without realizing the high-consequences.<br />
<br />
Users are unreliable security-partners to evaluate asymmetric threats.<br />
<br />
Users tend to dismiss potential problems if they are perceived as low-probability events. Organizations, on the other hand, have to look more carefully also at <u>what is at stake</u> and the <u>expected loss per event</u>, which can capture the high-consequence risk and motivate adequate counter-measures.<br />
<br />
Also important to many organizations today, for regulatory compliance with HIPAA and privacy regulations, are the potential high fines and breach notification duties that can be imposed, not just direct losses. <br />
<br />
<b>Usability and Security Aspects</b><br />
<br />
Adding a conventional security system to better counter asymmetric threats and provide regulatory compliance, will most likely require users to change. <br />
<br />
For example, if you must send to each recipient a password to allow reading the email, this makes it hard to use and is not natural. Or, if you must require recipients to register and select a password, just to read your email, this forces recipients to use yet another service they did not choose. Alternatively, if the solution requires going through a particular interface for webmail and/or install plugins for a desktop mail client, then users have to change their work environment. <br />
<br />
Further, requiring users to change burdens users with new procedures and disrupts use when desktop updates and plugins clash, reducing productivity. This can also end up blocking cost-saving and desirable options, such as Google Apps and phones, if they are not protected.<br />
<br />
Thus, notwithstanding the security and privacy needs, usability is also of critical consideration. Making it harder to use in order to be secure is not secure (either will not be used or be used poorly). Moreover, organizations have various non-compliant desktop, cloud, and phone systems that are in service and people already know how to use. There are other significant limitations in practice, such as non-compliant systems used by partners and customers, and customer support for new applications.<br />
<br />
<b>Users Do Not Want Change</b><br />
<br />
If there is any unanimity in what users want, it is that they want to use their systems without change! <br />
<br />
Second, users want to be able to switch to cloud or phone if they are not in the office, or if the office system is down. Third, users also want to communicate with their customers and partners without asking them to change. Fourth, users view security as "that which reduces productivity" and so will try to bypass security rather than follow a security process — for example, users will likely close a warning notice without even reading it.<br />
<br />
<i>The IT challenge today is how to provide the functionality that users want with the security that the organization needs.</i><br />
<br />
<b>The ZSentry Solution</b><br />
<br />
To maximize return on investment, organizations should invest first in areas that support multiple objectives within a diverse spectrum of asymmetric threats. The goal is to effectively enable security and privacy compliance while assuring usability and versatility anywhere, for all systems that people can or already know how to use. <br />
<br />
ZSentry is unique in providing these capabilities, through the various <a href="http://zsentry.com/setup.htm">Use Options >></a> that can be used in separate or concurrently, including ZSentry On-Site (desktop Mail clients), ZSentry Cloud (FISMA compliant), ZSentry App (web browsers), ZSentry API (custom services), and ZSentry SMS (secure text messages for phones).<br />
<br />
And given that no one is likely to successfully deter, prevent, or preempt every attack, organizations must invest in capabilities to help eliminate or at least mitigate the effects of a potentially catastrophic attack. <br />
<br />
ZSentry is unique also in allaying access and data security concerns locally, in the infrastructure, and in the cloud. Each customer's login and data are protected in separate by the Sans Target ZSentry technology, with configurable, encrypted metadata (keys are also protected by the Sans Target ZSentry technology) providing a protected, standards-compliant, unique user experience and feature set for each customer. <br />
<br />
This is even more important in the context of "cloud computing" and SaaS (Software as a Service), when user data may be stored in the "cloud". With ZSentry, customer access audit trails and customer data storage can be securely maintained in the "cloud" with encrypted, de-identified numbers, which access keys are provided and secured by the ZSentry technology.<br />
<br />
Read more at <a href="http://zsentry.com/setup.htm">Use Options >></a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-72038426388998554832010-11-26T20:13:00.000-08:002010-11-27T11:13:01.319-08:00Inclusive Specialization with HIPAA for Desktop and CloudWith or without HIPAA compliance needs, your organization is likely facing two clear choices today: Desktop or Cloud.<br />
<br />
The Desktop choice is interesting for business users, who commonly prefer to have their data local for privacy and control. In addition, Desktop systems such as Outlook are much easier for corporate setup and dealing with moderate to high mail volume, incoming or outgoing. And you can integrate data from different applications and different sources on the Desktop in ways that you cannot do with Cloud based solutions, such as in sending secure personalized messages merging each recipient's name and records.<br />
<br />
On the other hand, with the Cloud choice, well-known systems such as Google Apps, Gmail, and Yahoo, offer easy access from anywhere, much lower cost (even free), 24/7 maintenance, and other benefits. But Google Apps, Gmail, Yahoo, and other Cloud systems are not HIPAA-compliant.<br />
<br />
The Cloud choice privacy problem is solved by ZSentry, which enables Google Apps, Gmail, Yahoo, and other Cloud systems to be HIPAA-compliant. This allows the Cloud to be a good choice for Desktop replacement also in terms of HIPAA and privacy regulation compliance.<br />
<br />
However, because each choice has good points (otherwise, would not be a choice), choosing also means losing.<br />
<br />
This problem is also solved by ZSentry, which offers the <a href="http://zsentry.com/setup.htm#onsite">On-Site setup</a>, an inclusive specialization approach that works and is HIPAA-compliant for <i>both</i> Desktop and Cloud systems, as well as Web and Mobile systems. <br />
<br />
With ZSentry there is no need to choose, and lose. Based on metrics that are important to <i>your</i> case, each choice can be specialized to areas where it performs best, such as in terms of cost, usability, and security. The same applies to email, webmail, IM, SMS, storage and other communication choices. <br />
<br />
Rather than exclude valuable choices, the ZSentry <a href="http://zsentry.com/setup.htm">Setup choices</a> allow you to use each one where each performs best according to your metrics. And use the benefits of the Cloud platform also with HIPAA-compliant messages.<br />
<br />
HIPAA-compliant Desktop and Cloud use is further discussed in the ZSentry <a href="http://zsentry.com/setup.htm#onsite">On-Site</a> use option, and you can try it <a href="http://zsentry.com/freetrial.html">without cost</a> to see how it works for you. Other use options <a href="http://zsentry.com/setup.htm">are available</a>, including a HIPPA-compliant Cloud setup for Google Apps, that only uses a Web Browser.<br />
<br />
Comments are welcome.Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com2tag:blogger.com,1999:blog-17329179.post-59351937053833632992010-10-21T12:47:00.000-07:002010-10-21T13:30:22.666-07:00Identity Theft vs Impersonation FraudIf identity theft is a crime, mustn't identity be a property?<br />
<br />
No. And, clearly, this should show that there is no such as a thing as "identity theft". <br />
<br />
But why people, including laws in Congress, use <b>Identity Theft</b> instead of the long-established and legal term <b>Impersonation Fraud</b>? What's happening here? Was this just a change of fashion? <br />
<br />
It's because language is being used to shift the burden to the consumer. Using the soundbite "identity theft", instead of "impersonation fraud", redirects the loss to the victim and frees those actually responsible from any legally relevant questions that could be asked about their loose security of customer's accounts.<br />
<br />
"Impersonation fraud" as a term focuses attention on the fact that the criminal is deceiving someone in order to gain advantage by claiming to have some valuable characteristics or authorizations in fact belonging not to the criminal but to some other person. The person deceived is the primary victim in contemplation when this terminology is used.<br />
<br />
"Identity theft", by contrast, suggests that the victim is the person impersonated, because his or her "identity" has been "stolen".<br />
<br />
This way of looking at things implies that the losses which arise out of the impersonation fall on the person impersonated, rather than on the person deceived by the impersonation.<br />
<br />
"Identity theft" as a label is attractive to, for example, some banks who may wish to suggest that losses must be carried by their customers because they failed to take proper care of their "identity".<br />
<br />
Indeed, there is no such thing as an "identity theft" and there should be no responsibility to you if someone else uses your name and you do not know about it, in the same way that there's no responsibility to you if someone sends spam or mails a letter on your name.<br />
<br />
The burden for preventing the real crime of <b>impersonation fraud</b> should fall on those who accept the false identity and also on those who do not protect private records of their customers. Certainly not the consumer.<br />
<br />
If you feel that it is time to see real solutions to this problem, rather than let the discussion continue to be dominated by a special choice of language, What can you do?<br />
<br />
Let's look at five key points that you should think of:<br />
<br />
1. Think it through a bit more before considering to "buy protection". In fact, as discussed here, we should refuse to buy such "protection" and, instead request that those holding or using our identifiers follow the law and be held responsible for their misuse (as now enforced in the US with medical records) -- not the victims.<br />
<br />
2. Note that the use of the term "identity theft" by an organization trying to sell "protection" against it to the very victims should be illegal. At the very least, it should alert us to that organization selling a service and then, in case of loss caused by that service, trying to pass both the blame and the loss to the victims.<br />
<br />
3. The fact that people got used to talk in dumbed down soundbites such as "identity theft", instead of using well-established words like "impersonation fraud", does not mean that any legally relevant conclusions can be drawn from the misuse of technical terms like "theft" in the soundbite (as might be wrongly asked: If identity theft is a crime, mustn't identity be a property?).<br />
<br />
4. Keep in mind that "identity theft" is not a theft, and the victim may not even have been remotely at fault, as in terms of negligence.<br />
<br />
5. Do not allow loose language to redirect the focus elsewhere but where the problem really lies.<br />
<br />
Now, how can you protect yourself against "identity theft"? What if your information falls into the wrong hands?<br />
<br />
We increasingly communicate online instead of by telephone. Email, SMS and IM are the psychological equivalent of a voice conversation. People often forget, however, that a text message or an email conversation is stored online in many different servers and can be replayed many years later, often without the authorization of any of the parties, and often with undesired consequences.<br />
<br />
According to a 2010 survey commissioned by Microsoft, 90 percent of the general population and senior business leaders are concerned about the privacy and security of their personal data in the cloud.<br />
<br />
With ZSentry, you can actually eliminate the disclosure risk of email, webmail, SMS and IM by properly setting your message to self-destruct. This includes a combination of technical and legal measures, as done by secure email Zmail provided by ZSentry, and works with all editions.<br />
<br />
Read more at <a href="http://zsentry.com/encrypt-and-self-destruct.htm">http://zsentry.com/encrypt-and-self-destruct.htm</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-45479121866131697682010-09-27T13:39:00.000-07:002010-10-05T13:40:04.351-07:00ZSentry Premium launch<p>Anyone who was an early adopter of PKI (Public Key Infrastructure), PGP (Pretty Good Privacy) and other email security solutions will recall how difficult it was to explain how to send and receive secure email.</p><p>Secure email was one of those things that you couldn't really explain to people. It was something that senders and recipients had to see in action, something that they both had to learn and experience to really appreciate the way the technology would make their email secure.</p><p>Sending and receiving email is also one of those experiences that people just don't want to disrupt, not even to make it secure.<p><p>However, people more and more want to use email, webmail, SMS, and IM, and store documents online for easier access, while they also need to comply with HIPAA and other privacy regulations.</p><p>That's all solved with the commercial launch of ZSentry Premium in Oct/2010.</p><p>Since 2004, ZSentry has been in continuous use as a beta service and has served millions of secure email messages worldwide. ZSentry provides the needed security but does not change the way people use email or anything else, and works with desktop solutions such as Outlook as well as Gmail and Google Apps without plugins.</p><p>A fully-functional free version, called ZSentry Basic, will also continue to be available for personal use.</p><p>To see how it works, please explore the <a href="http://zsentry.com/zsentry-how-to.htm">ZSentry How-To</a></p><p>While ZSentry can be provided on-site, it provides a convenient "pay-as-you-go" cloud service, where it innovates by <i>eliminating</i> critical security and privacy risks associated with cloud computing. Each customer's data is protected in separate by ZSentry's "Sans Target" ZSentry technology, with configurable, encrypted metadata (keys also protected by the "Sans Target" ZSentry technology) providing a protected, standards-compliant, unique user experience and feature set for each customer.</p><p>With its innovative design, ZSentry helps allay data storage security concerns, both locally and in the infrastructure. This is even more important in the context of "cloud computing" and SaaS (Software-as-a-Service), when user data may be stored in the "cloud". With ZSentry, customer access audit trails and customer data storage can be securely maintained in the "cloud" with encrypted, de-identified numbers, which access keys are provided and secured by the ZSentry technology.</p><p><a href="http://zsentry.com/freetrial.html">Free Trial</a> and other options are available.</p>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-20053939859205272502010-08-21T11:26:00.000-07:002010-10-05T12:15:22.242-07:00Why you need to self-destruct your emailEmail, SMS and IM are the psychological equivalent of a voice conversation.<br />
<br />
People often forget, however, that a text message or an email conversation is stored online in many different servers and can be replayed many years later, even without the authorization of any of the parties, and likely with undesired consequences.<br />
<br />
The Internet "cloud", with webmail and more online processing services, just increases the privacy risks. According to a 2010 survey commissioned by Microsoft, 90 percent of the general population and senior business leaders are concerned about the privacy and security of their personal data in the cloud.<br />
<br />
Your data privacy rights also vary in online versus local use (e.g., on an individual's USB token, or hard drive in the home or office), as well as "in transit" versus stored. Anyone familiar with the privacy limitations of current laws written in the pre-Internet era, such as the 1986 U.S. Electronic Communications Privacy Act, knows that the individual's legal rights to privacy and protection from unreasonable search are starkly reduced for information stored online. Data that individuals store online, in cloud computing services such as webmail, receives a much lower level of privacy protection than data "in transit" or stored locally.<br />
<br />
Therefore, even though email may be the legal equivalent of a written conversation, email is mostly not legally protected online. That's one important reason to limit email liability. Other reasons include preventing harassment, coercion, blackmail, or just plain embarrassment.<br />
<br />
But, how? An email message may live in many clients, servers, and repositories, some of which may be covert and unreachable by you and your ISP.<br />
<br />
Surprisingly as it may seem, you can actually <i>eliminate</i> the disclosure risk of email by <u>properly</u> setting your email to self-destruct. This includes a combination of technical and legal measures, as done by secure email Zmail provided by <a href="http://zsentry.com">ZSentry</a>.<br />
<br />
This is how it works.<br />
<br />
ZSentry encrypts and self-destructs (expires, with no action on your part) your webmail, email, SMS or IM message, so that the intended recipients are allowed to read it only within its retention time, and neither you nor the recipients are responsible or liable for destroying it after it expires.<br />
<br />
The command to expire is clearly noted and also provided with US and international legal support by notifying the reader that the message is copyrighted and that the sender only allows reading during its retention time. Therefore, if someone wants for example to take a picture of the message, then using that picture after the expiration could be considered a breach of copyright and illegal circumvention of protection. While the first motivation is to reduce exposure, the second is to provide a legal recourse in case of exposure.<br />
<br />
After expiration there is no risk, as keys are deleted and any physical copy is therefore erased and not recoverable. Because self-destruct happens after a point in time that was known beforehand by all parties, claims of intentional destruction should be void.<br />
<br />
What about before expiration? With ZSentry, you can request a Return Receipt so that any attempt to breach access security of your zmail is immediately logged and traced, and you can be notified as well by a Return Receipt that is sent to you with full access information including Who, Where, When, and How.<br />
<br />
More information at <a href="http://zsentry.com/how_zmail.htm">How Zmail Works<br />
</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-70789271305895129262010-07-21T09:35:00.000-07:002010-07-21T18:46:06.574-07:00Spam, Spoofing, Phishing, PharmingAre you concerned about email fraud, how to keep your message private, and avoiding identity theft? Do you think that people receiving an email attachment that was clearly sent from you (as stated in the From line) may fear opening it?<br />
<br />
I frequently receive email from myself, that is with my name in the From line, that I never sent. It's well-known that anyone can send a regular email using your name and email address. This is used in spam attacks.<br />
<br />
But, why should you care? <i>"After all, it's just an email and I can simply delete it."</i> <br />
<br />
However, hidden persuaders in a message (or just in its Subject) may trigger your inadvertent reply to it, such as hearing that you are now listed in "The Internet Harvard Who's Who" (a fake honor), that your site can get a free evaluation (a scam offer), or an "alert" that someone is registering the left-most label of your .COM domain name in the Taiwan registry under .TW (which is legal and should not affect you). Or, and even more effective for the attacker, that triggers an irrational fear factor, where you fear things you shouldn't and then put yourself in greater danger. For example, a message from your bank, looking as the "real thing" and asking you to login to change your password.<br />
<br />
So, as in many other cases, an initial, simple attack that is not checked can lead to secondary attacks with far more damaging consequences than originally thought possible. <br />
<br />
What's gaining in importance in this space, although much less well-known than spam, is that attackers often use spam as a vector for delivering secondary attacks.<br />
<br />
The most relevant secondary attacks today include not only spoofing, phishing, and pharming (see GLOSSARY), but also installing a "trojan horse", a computer virus, or crippling your computer with a ransom request to fix it.<br />
<br />
Of these attacks, the most insidious and prevalent today is phishing, which can lure recipients to disclose their private data, that criminals worldwide (not necessarily just the attacker) can use to mount a real-world threat to your identity, bank account, and business, leading to potentially large losses. And criminals can do it all from an Internet cafe, somewhere in a far, far way place, or even next to you, and you (or the FBI) will not likely ever be able to hold anyone accountable for it.<br />
<br />
How did we get to this point? Some think that the reason is that your email address is global and even searchable; that's why your mailbox is overflowing with spam. In short, one is led to think that there is no way to prevent it. You must fight it by purchasing solution X (a spam/spoofing/phishing filter), which will require frequent paid updates to be effective. But such solutions can only protect you against yesterday's attacks, and even that may fail as attacks have many variants.<br />
<blockquote><i>Do I need to change my email address, Mail app, or provider, to be secure?</i> No, the problems are not due to your email address or Mail app, but how you send and receive email. You should be able to use any email address, Mail app, and provider you want, including webmail.<br />
<br />
<i>How about if I add a firewall and always use SSL?</i> Still, solution X (a spam/spoofing/phishing filter) will be killing good emails, and you really shouldn't open attachments even if you know the sender.</blockquote>In addition, anyone can read any email that you send and receive, so that you cannot use regular email in HIPAA and regulatory privacy compliance. Regular email communication is simply not secure. Sending an email is similar to sending a postcard. Any regular email that is sent by you or to you may be copied, held and changed by various computers it passes through, as it goes from you or to you. Persons not participating in your email communications may intercept your communications by improperly accessing your computer or other computers, even some computer unconnected to any party in the communication, which the email passed or was made to pass through. That's also at the root of the spam problem. In the same way that anyone can send a postcard in your name, anyone can send a regular email using your name and email address.<br />
<br />
NMA has developed ZSentry to protect your email against spam, spoofing and phishing emails, while adding a number of security and usability features that are missing in email, <i>without changing your email address, Mail app, or provider</i>. ZSentry also encrypts your email per-message, providing HIPAA and HITECH Safe Harbor compliance, as well as compliance with other privacy rules, with no Business Associate Agreement to be signed.<br />
<br />
How does this work against spam, spoofing, phishing, and pharming? Rather than fight them, ZSentry prevents them by (among other features): <ol><li>authenticating the <u>source</u> (including the sender's location) of a message; and</li>
<li>authenticating the <u>name</u> and <u>email address</u> of senders and recipients.</li>
</ol>For example, if a Zmail (ZSentry Mail) comes to you from the email address <friend@isp.com>, and you can decrypt it using the ZSentry service, then you have strong, cryptographic evidence that it did come from that address as cryptographically authenticated during signup, with the original subject, date, body and attachment intact.<br />
<br />
To read a Zmail, ZSentry also reminds users that they can copy-and-paste the ZSentry link (when, selected by the sender, Zmail uses an encrypted link). Rather than just click on the link, this simple procedure prevents users from landing at a destination that was encoded in the email to be different from what users can read on the screen before they click. <br />
<br />
Users can also beneficially apply a spam/spoofing/phishing filter prior to reading the Zmail, to reduce input email volume. However, the filter no longer has a critical, final function. This also means that the filter does not have to be set so tight as to increase too much the number of false positives (the number of rejected but good email), or should one fear that the filter is letting through too many false negatives (the number of accepted but bad email).<br />
<br />
A further benefit of the ZSentry approach is that it does not require the customer to update anything in order to remain protected.<br />
<br />
Other approaches (solution X, a spam/spoofing/phishing filter) such as those based on email headers, reputation, non-verifiable metrics (eg, community detection), blacklists, pattern detection, heuristics, zombie detection, and message scanning, can break privacy and may easily fail. One of the reasons to fail is that spam and phishing emails are created in an arms race scenario, where defenders lag behind with less knowledge and resources and are often fighting (perhaps, even well) the last exploit but not the next. Exploits are also hard to filter because they have many variants, spoof various parts of email headers and body, and come in the name of people or organizations you trust — your friends and business contacts. You probably receive several emails from yourself (surely, it is a valid email address and one that belongs to a real, reputable person) that you never sent.<br />
<br />
Training users to detect spam, spoofing and phishing adds costs and also frequently fails, as users cannot be trusted to follow procedures, are easily distracted, and may not understand the instructions in the first place. ZSentry does <i>not</i> depend on users learning ever-changing patterns, as this is one of the few things that is actually proven not to work, or just work poorly.<br />
<br />
How about spam? ZSentry also has a zero-tolerance spam policy. There are several mechanisms in place to prevent any ZSentry user from abusing the system and sending Zmail spam. For example, ZSentry BASIC users can send a limited number of secure email Zmail messages a day. ZSentry PREMIUM users, who must provide a valid payment information and physical address in order to use the service, are allowed to send larger amounts of secure email.<br />
<br />
<b>GLOSSARY</b><br />
<br />
<b>What is a "spoof web site"?</b><br />
<br />
A spoof website is one that mimics another website to lure you into disclosing confidential information. This can be done even with SSL (Secure Sockets Layer) using 128-bit encryption. To make spoof sites seem legitimate, spoof web sites use the names, logos, graphics and even code of the real company's site. They can even fake the https web address that appears in the address field at the top of your browser window and the "SSL padlock" that appears in the lower right corner of your browser.<br />
<br />
<b>What is a "spoof email"?</b><br />
<br />
A spoof email has the "From:" header of the email, and possibly other headers as well, set to the email address of a different sender, to lure the recipient to read and act on the email. For example, using the email address of a friend, a legitimate company, a bank or a government agency. This is very easy to do with regular email. To make spoof emails seem legitimate, the email body uses the names, logos, graphics and even legitimate web addresses and email addresses in some fields. The action links in the spoof e-mails almost always take you to a spoof web site. Spoof emails can be sent also as an attack against you or your organization, with fraudulent offers, bogus announcements or malicious content.<br />
<br />
<b>What is a "phishing email"?</b><br />
<br />
Phishing (or hoax) emails appear to be from a well-known company but can put you at risk. Although they can be difficult to spot, they generally ask you to click a link back to a spoof web site and provide, update or confirm sensitive personal information. To bait you, they may allude to an urgent or threatening condition concerning your account. Even if you don't provide what they ask for, simply clicking the link could subject you to background installations of key logging software or viruses. Every business on the Internet is a potential victim of phishing email attacks, eroding the trust of their customers in the company's communications.<br />
<br />
<b>What is "pharming"?</b><br />
<br />
A pharming attack redirects as many users as possible from the legitimate website they intend to visit and lead them to malicious ones, without the users' knowledge or consent. A malicious site can look exactly the same as the genuine site. But when users enter their login name and password, the information is captured. Emailed viruses that rewrite local host files on individual PCs, and DNS poising have been used to conduct pharming attacks. Even if the user types the correct web address, the user can be directed to the false, malicious site.<br />
<br />
<b>What is "spam"?</b><br />
<br />
All Internet users should by now know about spam. The word spam as applied to email means Unsolicited Bulk Email. Unsolicited means that the recipient has not granted verifiable permission for the message to be sent. Bulk means that the message is sent as part of a larger collection of messages, all having substantially identical content. Usually, a message is spam if it is both Unsolicited and Bulk. Unsolicited email is usually normal email (examples include first contact inquiries, job inquiries, and sales inquiries). Bulk email is usually normal email (examples include subscriber newsletters, discussion lists, information lists, and update announcements).<br />
<br />
Comments are welcome.Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-50683750567032889332010-06-26T11:41:00.000-07:002010-10-23T23:16:35.662-07:00White House Seeks Comment on Trusted ID PlanIt's important to protect privacy! This is a comment on a federal draft plan calling for the U.S. government to work with private companies to create what the White House Blog calls <a href="http://www.whitehouse.gov/blog/2010/06/25/national-strategy-trusted-identities-cyberspace">The Identity Ecosystem</a> — an online environment <i>"where individuals, organizations, services, and devices can trust each other because authoritative sources establish and authenticate their digital identities,"</i> also called "The Trusted ID Plan".<br />
<br />
<b>Introduction</b><br />
<br />
Some have said that The Identity Ecosystem proposal is a call for a mandated Internet "driver's license". Others believe that The Identity Ecosystem would allow Internet users to complete transactions with confidence.<br />
<br />
Our opinion is that while it could be the former, it cannot be the latter as it stands.<br />
<br />
We are, thus, motivated to present this proposal, which supports the objectives of The Identity Ecosystem without adding any form of central control (or even fear of), while it allays privacy concerns, especially online. Further, it can be easily commercialized using an ecosystem with open participation, and internationalized, without any changes in current laws, which many privacy advocates argue do not sufficiently protect privacy online in the US and elsewhere.<br />
<br />
<b>How Did We Get Into This?</b><br />
<br />
Before we present our contribution, we would like to invite us all to step back and ask "How did we get into this?"<br />
<br />
Accounts may vary somewhat on how the Internet came to be what it is today. However, our exposition will not start controversially. We will stay on undisputed facts in terms of authority and authoritative sources in the Internet, which is quite central to the The Identity Ecosystem proposal.<br />
<br />
Starting with the transfer of responsibility and authority for Internet policy increasingly from DARPA to NSF and to the Federal Networking Council after about 1988, NSF's role increased substantially with the creation of the NSFNET.<br />
<br />
At that time, central control of the Internet as initially provided by DARPA was effective, as evidenced by the fact that spam was famously not allowed and did not exist.<br />
<br />
But as the transfer process from DARPA continued, the central control paradigm started to change and weaken. Other points can also be mentioned, such as the role played by BBN in showing that it could work, and the gradual decrease of NSF's role to the point were the NSF was prohibited from spending money on it. And, on October 24, 1995, the Federal Networking Council famously passed a resolution defining the term <i>Internet</i>.<br />
<br />
Of course, control did not evaporate immediately after 1988 or even 1995. As control was relaxed, fear of control remained for a long while, as shown in Dr. Jon Postel's famous failed attempt to relax control by redirecting Root to the IANA (Internet Assigned Numbers Authority) on Jan. 28, 1998.<br />
<br />
This is the big picture that matters now, in 2010. From 1988 onwards, the US Government has deliberately relaxed its control over policy as the Internet has become increasingly commercialized and internationalized. Hence, gradually, confidence has evaporated because it was based on that, now missing, central control.<br />
<br />
This is how we got to be where we are, by the very nature of growing into and being a network of networks, where no one can control both ends of a general communication channel (neither sending nor receiving).<br />
<br />
<b>The Internet, Confidence, and Trust</b><br />
<br />
The Internet was born within a strict central control system (DARPA), but then it grew and now (as we saw above) abhors central control. This development was enabled by technology, allowing networks of networks to be easily built and expand. However, it is not unlike what we find elsewhere. It seems to be a natural process, and not one that we can reverse (or should try to reverse).<br />
<br />
If central control cannot be restored, how can confidence be restored? How can rogue operators be detected and prevented from using protected resources?<br />
<br />
The word confidence means "with faith". It is a good concept to use when we can have an authoritative source (e.g., DARPA). But fails miserably when such source does not, or cannot exist.<br />
<br />
In such cases, we can use a concept called trust that has been intuitively known for thousands of years. Trust has been formally defined (Gerck, 1998) in terms of Information Theory, as "trust is that which is essential to a communication channel but cannot be transferred from a source to a destination using that channel".<br />
<br />
This is an implicit definition, allowing us to derive many equivalent positive and negative statements and use them as "domain definitions", so that we can use the concept of trust coherently through different domains including with people and machines. See <a href="http://bit.ly/TRUST">http://bit.ly/TRUST</a> and <a href="http://bit.ly/IT-TRUST">http://bit.ly/IT-TRUST</a><br />
<br />
For example, the definition implies that a decision to trust someone, the source of a communication, the name on a certificate, or a record must be based on factors outside the assertion of trustworthiness that the entity makes for itself.<br />
<br />
As another example, in terms of a conversation, we can use the equivalent domain definition (see references) "trust is that which provides meaning to information". Indeed, according to the trust we have on the information source, the same words may have quite different meanings -- as it can be easily demonstrated by watching the different meanings presented in newscasts by FoxNews and MSNBC, for the very same information.<br />
<br />
The important point here is that trust can be based on other factors, in addition to control or even fear of control.<br />
<br />
<b>Solution Considerations</b><br />
<br />
A solution to the perceived identity problem in the Internet should, thus, investigate what <i>other</i> factors must now be introduced in order for trust to be induced without trying --and failing-- to re-introduce control, or fear of control.<br />
<br />
In other words, there's no "Doomsday" scenario of control vs anarchy either way. Rather, there are many reasons to abandon the mindset of recourse to control (or fear of) as the only solution.<br />
<br />
For example, we need to take into account that the Internet is essentially <i>open-ended</i>. Anyone can join some network connected to networks comprising the Internet. The Internet is a network of networks, with no common reporting of any kind that could allow objective authorizations to be authoritatively defined and enforced, for all the different networks in the Internet.<br />
<br />
We also share the same network as the attackers. Even if all identities were somehow centrally controlled, the false use of identities would not be preventable (as we know well from the brick-and-mortar world, for example). This further indicates that we shall not hope to be able to confine what we cannot control.<br />
<br />
Users, furthermore, are bound in their defenses by their own usability and resource limitations, and service providers must only deploy defenses that users may tolerate. This creates an "arms race" scenario with a large information and power asymmetry against users. Attackers know more and are much more powerful than users and even service providers, and can rather easily mount massive campaigns with massive resources (e.g., a phishing attack using thousands of bots), while defenders are frequently one step behind with the defenses they can deploy.<br />
<br />
It would, therefore, be better for users that we avoid an "arms race" scenario, where defenders (users) will lag behind. Instead, we motivate the need to find ways and "methods of peace" in achieving Internet security, where security must be primarily viewed as a form of understanding, not of confinement or subjugation. For example, it makes little sense to confine use to X, if one has no justification to trust that which one is using in order to confine use to X or, if one does not know what one is confining in or out.<br />
<br />
Thus, unless we are conditioned to think that the only way to effect security shall be by subjugation and fighting the "thing" out to the bitter end, which will mean the defeat of users and their rights, we should not pursue control (or fear of) as the only solution. It does not fit the problem and does not help users.<br />
<br />
And the main visible point to users should be about <u>how to make non-conformance public</u> rather than <u>certifying conformance</u>.<br />
<br />
Not only would there be then much less liability for the service, but the user is kept in the verification loop --as the user should be-- rather than blindly rely on some sort of oracle. Also, in security terms, not only less attacks are possible but attacks are so not direct in creating an error condition. This does not mean that the user would manually have to make the determination in every case and be burdened by it. The non-conformance detection can be automated, classified by risk, cached, and previous results can used until they expire according to their respective risk class.<br />
<br />
Along these lines, identity should not only refer to an attribute (name) or some aggregation of attributes (name, address, SSN) but, even more strongly and effectively refer to relationships ("Is there an identity relationship between the true owner of this Mastercard and the person presenting it to me?"), connections, absence of connections, and other connection properties. It's not about one authoritative channel either, to vouch for the identity, but using multiple factors to prevent a single point of failure.<br />
<br />
<b>Privacy Rights</b><br />
<br />
Let us think about the US Government getting involved in assuring identity and protecting privacy at the same time. I'd like to mention Judge Scalia's private (but not secret) comment to a previous team member, who had an opportunity to speak with Justice Scalia in 1999, and asked him if privacy warranted strong legal protection in the US. Of course, this would be at least tangentially relevant to any US Government identity initiative such as the Trusted ID Plan.<br />
<br />
"Not based on the constitution", Scalia said. "Nowhere does the constitution address the right to privacy. If people want the Right for Privacy, they will have to change the constitution."<br />
<br />
More importantly, Scalia added that "Free Speech is defended by the constitution, and that makes the Right for Privacy very difficult to defend."<br />
<br />
The Scalia quotes are as verbatim as memory can allow, are not jurisprudence, and other Justices may disagree, but the idea is quite easy to verify — the US Constitution does not support the Supreme Court to rule in favor of privacy rights purely on constitutional grounds and could deny privacy rights in the event of a clash with the well-established constitutional right to freedom of expression (by a person or company).<br />
<br />
In other words, it is possible that a company's right to freedom of expression (as in the recent election spending Supreme Court decision) would trump an individual's right to privacy.<br />
<br />
It is further well-known that users' online privacy rights are reduced in the US by the 1986 Electronic Communications Privacy Act (ECPA), which is complex and often unclear. What is clear is that the ECPA does not mandate a search warrant to access online private communications and the locations of mobile devices. Data or files stored in the Internet "cloud," such as webmail or identity data, are subject to an often lower privacy standard than the standard that applies when the same data is stored on an individual's hard drive at home or in the office.<br />
<br />
Considering also the lack of privacy protection in many US States' constitution (except, for example, California and Washington as examples where privacy is strongly protected by their constitutions), and in the US Constitution, in addition to the well-known information asymmetry problems and other aspects (e.g., power asymmetry) that are behind the increasing market failure to protect users' privacy in the realm of private companies, as well as lack of international harmonization in this regard, the conclusion is clear:<br />
<br />
<i>Currently, we should be careful in relying on government, market (private companies), or legal recourse to protect users' identities.</i><br />
<br />
Thus, a <i>missing</i> and yet important part of The Identity Ecosystem proposal in this matter should be on helping to improve laws regarding identity and privacy, which will facilitate the adoption of useful technologies. <br />
<br />
<b>How about The Identity Ecosystem?</b><br />
<br />
Based on the discussion above, both technically and in terms of privacy protection, it will <b>not</b> work to have the Internet fall back to the idea of confidence supported by authoritative sources that would purportedly "establish and authenticate users' digital identities". Central control, and fear of it, is both not possible and not privacy-preserving.<br />
<br />
Thus, if we think of it in terms of a "Trusted ID Plan" supported by authoritative sources, it will be restricted in use to those authorities each user can trust (and, likely, pay) within their own desired extent, which varies for each user, each case, and also over time and location. This will likely lead to a number of disjoint local systems that will not be able to talk to each other or internationally.<br />
<br />
<i>However, The Identity Ecosystem may use a more comprehensive approach.</i><br />
<br />
Many real solutions are possible, but they should all be founded on the idea that trust can be based on other factors, in addition to control or even fear of control. For example, a user should be able to effectively refer to relationships or just even prior communication established online with other entities, in defining an assertion of identity. A solution should, thus, investigate what other factors must be introduced in order for trust to be induced.<br />
<br />
In so doing, we point out that self-assertions cannot induce trust (<a href="http://bit.ly/IT-TRUST">http://bit.ly/IT-TRUST</a>). Saying "trust me" should not make you trust me. The verifier needs to have access to multiple and varied sources of trust channels.<br />
<br />
Further, to be usable, we want to reduce user frustration in having to use a different tool if one needs security and regulation compliance. This also would help reduce the focus on security, so that people can long at last focus on what they want to do, not how they have to do it.<br />
<br />
<b>A Definite Proposal</b><br />
<br />
In addition to the considerations and suggested guidelines above, I also submit a definite proposal, that has been field-tested in millions of cases in the US and worldwide since 2004, and can be demonstrated today, to anyone, with zero cost.<br />
<br />
The author, based on the ideas of trust developed earlier (Gerck, 1998, op. cit), has developed a protocol that fits the above considerations, provides multiple and varied sources of trust channels, does not rely on "trust me", and yet provides a "no target" approach to eliminate concerns of attacks stealing identity credentials in online servers. It is the "identify & authorize" engine, as used by <a href="http://zsentry.com">ZSentry</a>.<br />
<br />
ZSentry is available <i>without cost</i> for personal use (Free Basic) and can be used with mail, webmail, file transfer & storage, IM, SMS, HTTP, fax and other communication protocols. ZSentry supports the ZS authentication mode, as well as PKI/X.509 and PGP. ZSentry also extends these standards in significant ways. An important issue solved, of course, is the problem of initial contact. ZSentry allows secure first contact and reply without previous interaction (e.g., exchanging passwords, requiring registration) or work (e.g., searching a directory, solving puzzles), and provides a number of life-cycle control functions, including release, expiration, and delivery verification.<br />
<br />
ZSentry also supports SAML and SSO, so that it can be part of a federated-identity ecosystem. And you do not have to worry about your identity being stolen online at the ZSentry site, or at other sites you visit. ZSentry uses its "no target" technology to protect your login credentials and keys at ZSentry, whereas the SAML identity authorization does not carry them either.<br />
<br />
Preventing false login (eg, by stealing a user's credentials with a key-logger) and duplicate use of the same account, may be a threat in some cases, especially with SSO. ZSentryID can be used to introduce a fresh second-channel challenge that changes for every authentication, for example by cell phone SMS. ZSentry also uses Adaptive Security, and other techniques to help allay such concerns.<br />
<br />
The trusted introducer function provided by ZSentry does not need to be carried over forever. This is <i>not</i> a single provider, lock-in proposal.<br />
<br />
Much like a booster rocket, once the transaction starts, other sources of trust are introduced (e.g., who do you know that I trust and can verify you by? What is your signed library key?) to the point that the ZSentry introducer function can be jettisoned without prejudice. <br />
<br />
With our proposal, there is no "trusted ID" that will suddenly lose all its evidence value if not renewed.<br />
<br />
Technical information on ZSentry identity verification is available at <a href="http://zsentry.com/identity.htm">http://zsentry.com/identity.htm</a><br />
<br />
I submit that ZSentry supports the objectives of The Identity Ecosystem without adding privacy concerns, especially online where it's not a matter of if but when and how often information on servers (even hosted at the Pentagon or FBI) <b>will</b> be disclosed.<br />
<br />
Best regards,<br />
Ed Gerck<br />
<br />
It's important to protect privacy! You can vote on our proposal if you wish at: <a href="http://www.nstic.ideascale.com/a/dtd/The-ZSentry-Proposal/45785-9351">http://www.nstic.ideascale.com/a/dtd/The-ZSentry-Proposal/45785-9351</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com10tag:blogger.com,1999:blog-17329179.post-60338003582268217812010-05-14T10:10:00.000-07:002010-05-14T11:02:22.704-07:00Cloud securitySome people say that cloud security comes down to this simple question:<br />
<br />
<i>Do you have 24/7 unrestricted access to the PHYSICAL machine where your data is stored? </i><br />
<br />
<i>If no, then your DATA IS NOT SECURE.<br />
If so, YOUR DATA MIGHT BE SECURE.</i><br />
<br />
This line of questioning goes to the heart of most security problems, and not just in cloud security terms.<br />
<br />
To further an example, rather than requiring a key to open a door (metaphorically, to access a data file), more secure systems would require 2, 3 or N keys such that a subset of M keys (for example, 3 out of 5, where N=5 and M=3) would be required to open the door. So, even if you lose the key in your pocket, or someone takes it from you under gun threat, you still have security within M of N. This is usually called a "security quorum" or "threshold system".<br />
<br />
Useful as a threshold system may be for cloud security, is there still room for improvement?<br />
<br />
A major improvement would be to <i>eliminate</i> the root of trust (i.e., that which you rely upon to assure that the threshold system and keys work as intended) [1] so that there is no target to attack. The principle is that no one can attack what does not exist. Not only a security quorum would not be needed (which can improve usability and prevent a denial of service attack) but also complexity would be reduced.<br />
<br />
This solution, which is implemented in NMA ZSentry [2], effectively shifts the information security solution space from the yet-unsolved security problem of protecting servers and clients against penetration attacks to a connection reliability problem that is easily solved today. <br />
<br />
Thus, in terms of cloud security, you can set up layers upon layers of security, but this is all for naught if the root of trust lies within the cloud itself. You solve this problem by eliminating the root of trust inside the cloud, as done by ZSentry, which is available free for tests and personal use [2].<br />
<br />
This is important not only for health care records for HIPAA compliance, and privacy regulatory compliance in general, but also for individuals. Email, SMS and IM are the psychological equivalent of a voice conversation. People often forget, however, that a text message or an email conversation in cloud storage (such as in gmail, yahoo, hotmail) can be replayed even many years later, often without the authorization of any of the parties, and often with undesired consequences.<br />
<br />
In perspective, while it may not be reasonable (or cost-effective) to assure that you have 24/7 unrestricted access to the PHYSICAL machine where your data is stored, you can use methods that deny the existence of the root of trust within the cloud itself. The cloud should only be used to store de-identified, encrypted data, and without the keys that would allow the data to be unraveled.<br />
<br />
<br />
[1] For the formal definition of trust, that applies to both computers and humans, see <a href="http://bit.ly/TRUST">http://bit.ly/TRUST</a><br />
<br />
[2] <a href="http://zsentry.com">http://zsentry.com</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-63303541217796216562010-04-26T17:10:00.000-07:002010-04-26T21:43:24.521-07:00Bank of America's SafePass Security: strike out(Posted online by Joel M Snyder. The problem and description are particularly relevant in our security and usability studies of access control systems.)<br />
<br />
Bank of America lets their patrons sign up for "SafePass," which is a credit-card-sized one-time password device. You MUST sign up for SafePass for certain transactions (like large transfers) but it is optional for most customers.<br />
<br />
I signed up. Much to my woe.<br />
<br />
The sign-up fee for the card is $20.<br />
<br />
I got my card, and it's physically defective: no number shows up when you push the button.<br />
<br />
To get this problem remedied, Bank of America has me in an infinite loop. The phone people cannot send me a new card (why? I don't know). So I am supposed to do this through the web site.<br />
<br />
However, once you have signed up for SafePass, you must use your SafePass to make any changes to SafePass (including getting a replacement card, and let's not even start on whether or not it's going to cost me another $20 to get a replacement for the first defective card).<br />
<br />
So I was transferred from customer service agent to customer service agent, and each one assured me that they need to get this card activated. Some thought that if they activate it on their end, through the miracle of the ether, this will suddenly cause numbers to show up on my display here. (Attempting to explain that this could not work turned out to be a losing battle). I went through supervisors. I went through supervisors to supervisors.<br />
<br />
In the end, the "solution" that they came up with--after 55 minutes on the phone, mind you--is that I should order a new card (not a replacement, which I cannot do, because the current card cannot be activated, but a new card, as if I need two of these things). They agreed to make a $20 credit on my account, and this will then add up to solution required.<br />
<br />
And all this because I was trying to do "the right thing..."<br />
<br />
jmsEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-68692610587254610532010-03-27T12:37:00.000-07:002010-03-27T13:10:38.367-07:00Red Flags In Email Security<small><i>REQUEST FOR COMMENTS: This article is based on a work draft submission by Ed Gerck, Ph.D. and Vernon Neppe, MD, Ph.D. -- your comments are welcome.</i></small><br />
<p>Solutions for email security are often plagued by hidden costs, hidden liability, and lack of both security and usability. Even though the choice of a security solution involves many aspects, what is clear is that a selection process should not overlook some obvious red flags.</p><a name="1"></a><p><b>Red Flag #1: Sign a HIPAA Business Associate Agreement. </b>This means that the security solution is exposing you to unnecessary technical, business, and legal risks.</p><blockquote><i>Signing a HIPAA Business Associate Agreement is mandatory if the solution does <big>NOT</big> take advantage of the HIPAA "Safe Harbor", which requires encrypted and de-identified data. In addition to legal model and harmonization problems for you in signing another contract, absence of the "Safe Harbor" means that if a breach of unsecured protected health information (PHI) is discovered, fines may be levied and notices to each person affected are required. In addition, contracts with a "Covered Entity" may be at risk and the reputation of all business involved will almost certainly be harmed.</i></blockquote><a name="2"></a><p><b>Red Flag #2: Centralized directory for email encryption.</b> Someone other than yourself has your business partners' and customer's contact information, with names, email addresses and other data.</p><blockquote><i>In addition to being a single point of failure, this means that your data may become available or be sold to solution providers and other businesses. This also means that in the event data is lost or stolen, it is linkable to a particular person, at risk of violating HIPAA, HITECH, and other regulations.</i></blockquote><a name="3"></a><p><b>Red Flag #3: It Just Works.</b> Beware of hidden liabilities. For example, make sure that your keys are not escrowed in the servers providing the solution, as with IBE (Voltage). Nothing is safe in servers, not only from attackers but also from service providers and employees.</p><a name="4"></a><p><b>Red Flag #4: Key Management "in the cloud".</b> Beware that nothing is safe "in the cloud".</p><blockquote><i>Customer data storage, including keys, can only be securely maintained in the "cloud" with encrypted, de-identified numbers, where the access keys should be provided and secured by yourself. See also comments to Red Flag #16.</i></blockquote><a name="5"></a><p><b>Red Flag #5: Service automatically detects and encrypts messages that contain personally identifiable information.</b> This means that your service provider actively stores and scans all your communications.</p><blockquote><i>Also sometimes mentioned as a "hosted solution that automatically encrypts email based on your policy definitions". This not only falls within the Red Flag #1 and excludes a "Safe Harbor", but potentially exposes your business to covert surveillance as well as to legal loopholes allowing service providers to use and sell your information as it is not "data in motion".</i></blockquote><a name="6"></a><p><b>Red Flag #6: Ensure Business Associates understand HIPAA implications.</b> This means that your business risk and reputation depends on others, who you cannot control. Instead, your security solution should limit the risk posed by others.</p><a name="7"></a><p><b>Red Flag #7: Keys provided by "Web-of-Trust".</b> This means that there is no entity responsible for revocation of keys, or for assuring that the keys are up-to-date, or valid for the intended purpose. This may work with a group of friends, but not when your risk and reputation depend on it.</p><a name="8"></a><p><b>Red Flag #8: Require Users to Buy a X.509/PKI or S/MIME Certificate. </b>For a variety of reasons, and not just cost, this has not worked since it was first proposed more than twenty years ago.</p><a name="9"></a><p><b>Red Flag #9: Protected with passwords. </b>Passwords are notoriously insecure and their management is a nightmare even for a few dozen users.</p><blockquote><i>Passwords are not considered secure by HIPAA, HITECH, the Federal Financial Institutions Examination Council (FFIEC), and other regulatory regimes. Compliant authentication solutions should use two-factor authentication, present familiar user interfaces (to prevent user errors), prevent dictionary attacks, and not store authentication credentials (such as passwords and usernames) online or, more desirably, anywhere.</i></blockquote><a name="10"></a><p><b>Red Flag #10: Use the fax. </b>It is a privacy and HIPAA violation to allow faxes with personal information to be available in an office area that can be accessed by anyone other than the intended recipient, which is also notoriously difficult to ensure.</p><a name="11"></a><p><b>Red Flag #11: Our employees are trusted. </b> Your business data security should not depend on trusting the service provider's employees ("we promise we won't look"), who you cannot select, verify or control. It is known that more than 70% of all security breaches are due to internal causes. Not even the FBI can prevent having a national security traitor for many years among their own directors.</p><a name="12"></a><p><b>Red Flag #12: We train your employees. </b> This means that your employees will need to be trusted to follow security procedures, which failure will fall on your business and is at the root of more than 70% of all breaches. See Red Flag #11.</p><a name="13"></a><p><b>Red Flag #13: Just install a plug-in for your Mail client or Web browser. </b>Also mentioned as "downloaded and installed in minutes". This means that you have to download an untrusted new component, that opens your system to new zero-day exploits, will need to be always updated, and may cease to work when you update your application or other plugins.</p><a name="14"></a><p><b>Red Flag #14: Delivered using a secure SSL/TLS encrypted connection. </b>This delivery process falls short of basic security requirements for email messages (see references below). See also Red Flags #1 and #5.</p><a name="15"></a><p><b>Red Flag #15: If the recipient is not registered, the message is sent to a secure portal. </b> This means that the recipient must register anyway, and reduces usability.</p><a name="16"></a><p><b>Red Flag #16: We can use your information for purposes including... </b> This statement may sound assuring but it is open-ended and does not exclude any purpose or person.</p><blockquote><i>This Red Flag is actually worse than the notorious "opt-out" policy, where sharing your protected information is the default mode, as there is no limit that you can request (by opting-out) regarding the use of your protected information.</i></blockquote><a name="17"></a><p><b>Red Flag #17: Beware of conflicts of interest between the distinct roles that may be played by the same service provider, which may increase both your risk and liability. </b> Also called the "fox taking care of the hens" scenario.</p><blockquote><i>For example, if (1) a service provider for secure email is also (2) a provider of anti-virus scanning service, then the service is potentially made aware of protected information and the "Safe Harbor" provision may not apply (see Red Flag #1).</i></blockquote><a name="18"></a><p><b>Red Flag #18: Impossible to break. Our servers are in a secure facility. </b> It is not a matter of if, but when.</p><blockquote><i>Regardless of how safe and secure people claim something online is, there always seem to be someone who can eventually crack access to it. Not even US Department of Defense and Pentagon servers are secure.<br />
<br />
Security claims should not depend on being "impossible to break" (the Fort Knox approach), but on the impossibility of finding anything of value when that happens. This strong concept of security does not assume any limitation on the attacker, which limitations are often found to be unrealistic but only after an attack succeeds.</i></blockquote><h2>Regulatory Compliance and Fines</h2><p>These and other Red Flags are important because security and email encryption are gaining importance, for some very clear reasons:</p><ul><li style="padding-bottom:10px">Email phishing, spam, email disclosure, and identity theft are major sources of fraud losses today.</li>
<li style="padding-bottom:10px">Protecting consumer privacy is becoming a duty for organizations worldwide.</li>
<li style="padding-bottom:10px">Organizations in regulatory privacy and security compliance regimes (e.g., HIPAA and FFIEC) need to communicate in a way that is compliant.</li>
<li style="padding-bottom:10px">After 02/2009, fines for non-compliance (HITECH Act) have sharply increased up to $1.5 million.</li>
</ul>For example, in the health-sector, service providers must comply with the US HIPAA (Health Insurance Portability and Accountability Act), HITECH Act (Health Information Technology for Economic and Clinical Health Act of 2009) and other regulations.<br />
<h2>Conclusions</h2>The choice of a security solution involves many aspects, including not to overlook the 18 red flags we discuss in this article. We also observe that privacy and security compliance is not enough — any security solution must be, first and foremost, usable. Otherwise, it will not be used or used incorrectly, defeating the security objective.<br />
<h2><i>References</i></h2>Gerck, E. (2007). Secure email technologies X.509/PKI, PGP, IBE and Zmail. In Corporate Email Management, Chapter 12, Edited by Krishna SJ, Raju E., pp.171-196, Hyderabad, India, ICFAI University Press. Available online at <a href="http://email-security.net/papers/pki-pgp-ibe-zmail.pdf" target="_blank">http://email-security.net/papers/pki-pgp-ibe-zmail.pdf</a>.<br />
<br />
Neppe, V. M. (2008). The email security-usability dichotomy: Necessary antinomy or potential synergism?. In Telicom, 21:3, May-June, pp.15-31. Available online at <a href="http://email-security.net/papers/usable-secure-email.pdf" target="_blank">http://email-security.net/papers/usable-secure-email.pdf</a>.<br />
<br />
Whitten, A. and Tygar, J. D. (1999). Why Johnny can't encrypt: A usability evaluation of PGP 5.0. In Proceedings of the 8th USENIX Security Symposium. Available online at <a href="http://www.gaudior.net/alma/johnny.pdf" target="_blank">http://www.gaudior.net/alma/johnny.pdf</a><br />
<br />
Feghhi, J., Feghhi J., and Williams, P. (1998). In Digital Certificates: Applied Internet Security. Addison-Wesley, ISBN 0-20-130980-7.</p>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-65146595780380775082010-02-24T20:45:00.000-08:002010-03-26T11:32:23.876-07:00Large EMR privacy breach notification, two years later -- an exception or a symptom?<i>NOTE: A colleague and I are working on a paper discussing a number of privacy and security red flags that can help call attention to these and other issues, especially in the context of secure email. A draft is gladly available to those who are interested, by private email request, for comments before publication. </i><br />
<i> </i> <br />
Electronic medical records (EMRs) are at the heart of health care reform, and there is both a personal as well as a legal expectation of privacy for EMRs. <br />
<br />
Promptly notifying users of privacy breaches can help bring accountability to the system, and help users.<br />
<br />
In February 2010, RelayHealth (also known as NDCHealth Corporation) acting as a claims processing company, notified prescription holders that EMRs of two years ago, dating between February 2008 and December 2008, with full name, date of birth, prescription number, insurance cardholder ID, and drug name, that were dispensed at Rite Aid as well as other retail chain pharmacies and independent pharmacies in the State of California, were sent to an unauthorized pharmacy.<br />
<br />
After I mentioned this case online, RelayHealth has contacted us in March 2010 and stated that the data was sent in error only in November 2009, so the delay in informing consumers was not two years but three months. <br />
<br />
However, we note that this information was not provided to RelayHealth's consumers when the privacy breach was disclosed in February 2010. RelayHealth may want to look into that communication and verify why they did not disclose a delay of three months. <br />
<br />
Further, what matters to our analysis here is the consumer privacy risk, which includes the two-year delay. If, for example, three-year old files are wrongly disclosed today and the EMR processor informs the patient tomorrow, this is not a lesser problem for the patient (as the next mentioned Fortis case exemplifies). <br />
<br />
The 2010 breach notification did not disclose why the information was sent (Who requested it? Under what authorization? Who approved it?), who incorrectly received the EMR, and who was responsible for the breach, neither what compensation or recourse users may have. <br />
<br />
In a recent court case, Fortis (a US health insurance company) was found to have a practice of targeting policyholders with HIV. A computer program and algorithm targeted every policyholder recently diagnosed with HIV for an automatic fraud investigation, as the company searched for any pretext to revoke their policy. <br />
<br />
Companies such as Fortis can find out about anyone's diagnosed HIV, or other illness, through pharmacies and claim processors, for example. <br />
<br />
This situation underscores the underlying conflicts of interest between at least three distinct roles that RelayHealth plays. They are: <br />
<ol><li>claims processor;</li>
<li>provider of patient EMR to their pharmacies and doctors;</li>
<li>provider/seller of EMR to providers other than the patient's. </li>
</ol>This last activity has the greatest potential conflict, as patients are included in a no-opt-out policy at <a class="moz-txt-link-abbreviated" href="http://www.relayhealth.com/">www.RelayHealth.com</a> that says (words in square brackets are comments, not from RelayHealth): <br />
<br />
<i>"Your Provider, a Provider-Designated User [pretty much anyone] or authorized member of a Provider Group [anyone] can use contact and/or health information about you stored by RelayHealth for many purposes including [ie, this says that it does not exclude anyone or anything]:</i><br />
<i>..." </i><br />
<br />
and <br />
<br />
<i>"RelayHealth may use the contact, billing and/or health information [EMR] provided by you in our service to provide your physician or other healthcare provider [ie, anyone they want] with updated and/or supplemental information for their files or systems." </i><br />
<i> </i> <br />
A pattern that seems to emerge here is that because EMRs also have a market value (for example, to insurance companies, pharmacies, etc.), health care service companies can build automated information exchanges where they can make collected EMR available to other entities, and build a business on this activity. <br />
<br />
That the same health care service companies (with different hats) also serve on behalf of the patients to protect the EMR from disclosure, is where the fox is taking care of the hens, and where the conflicts in 1-2-3 may play a role.<br />
<br />
What this means is that the expansion of health care into larger use of EMRs ought to call for a much broader review of procedures and conflicts of interest than what is currently available. And, obviously, it should also include stricter rules for information security and handling of EMRs than what's currently used. <br />
<br />
Your comments are welcome.<br />
<br />
Best regards, <br />
Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-41304306794626507992010-01-17T12:14:00.000-08:002010-02-01T12:56:25.210-08:00SSL would prevent it, Re: Internet security flaw exposes private dataWe all read about the most recent Internet security flaw that exposes private data [1]. Without any action on their part, a number of AT&T smartphone users found themselves logged into Facebook, a popular social networking site, under user accounts other than their own. The problem was quickly attributed to "misrouting," a term that suggests that information took a wrong turn somewhere in the network.<br />
<br />
It looks like AT&T did <i>something</i> wrong, but in terms of proxy server setup, not routing, and the company is in the process of fixing it. But Facebook should also share some concern with this, as it didn't consider the information tied to a user's account authentication to be important enough to protect with something stronger than clear text.<br />
<br />
So, when we look closer, the sky is not falling. <br />
<br />
With at least one authenticated end-point (the end-point server, as usual), SSL would NOT allow an AT&T caching server that sits in-between to have had a "correct" SSL session between the wrong two end-points. See A. Menezes et al., <i class="moz-txt-slash"><span class="moz-txt-tag">/</span>Handook of Applied Cryptography<span class="moz-txt-tag">/</span></i>, CRC Press, New York, 1997. <br />
<br />
This statement is true even if the AT&T caching server is able not only to modify packets but also to drop incoming packets and insert its own packets at will into the traffic, and do so in two-way communication without delay, becoming an active man-in-the-middle (which a passive, caching server is not). <br />
<br />
This should not be confused with a phishing attack, where the end-user is tricked to go to (for example) paypal.fraud.com instead of paypal.com -- and the server end-point is already wrong, or a "bridge" attack where (without noticing it) the user goes to the wrong server before SSL starts -- and stays there after SSL starts. The second user did reach facebook.com and this would have been enough under one-point authenticated SSL to prevent the first user session to be hijacked (which is what happened) as it would not have the correct session keys. <br />
<br />
This should also not be confused either with other attacks on privacy and security, that work in spite of SSL. For example, if SSL is used you may still be connected to the wrong site (eg, using a rogue certificate), and your traffic may still be read by others behind an SSL proxy, but you would not be connected to someone else's previous SSL session, not would your SSL session be usable by someone else.<br />
<br />
These attacks, despite much confusion online in discussion lists [for example, see refs. 2 and 3], do not break SSL nor have any bearing on the SSL being able to prevent the reported facebook case ("Internet security flaw exposes private data").<br />
<br />
The AT&T problem that originated this security issue -- which, by all evidence, was not an attack -- would have been prevented by facebook requiring SSL. This may be particularly important for mobile devices, <br />
where users may have less connection information available to verify.<br />
<br />
And, google did the right thing now by making SSL the default for gmail. <br />
<br />
REFERENCES: <br />
[1] <a href="http://arstechnica.com/web/news/2010/01/facebook-att-play-fast-and-loose-with-user-authentication.ars">http://arstechnica.com/web/news/2010/01/facebook-att-play-fast-and-loose-with-user-authentication.ars</a><br />
<br />
[2] <a href="http://www.listbox.com/member/archive/247/2010/01/sort/time_rev/page/4/entry/13:270/20100117153354:9FD73A7C-03A7-11DF-8606-4E77A52306B0/">http://www.listbox.com/member/archive/247/2010/01/sort/time_rev/page/4/entry/13:270/20100117153354:9FD73A7C-03A7-11DF-8606-4E77A52306B0/</a><br />
<br />
[3] <a href="http://www.listbox.com/member/archive/247/2010/01/sort/time_rev/page/4/entry/11:270/20100118113841:EE029CD4-044F-11DF-868D-868DA52306B0/">http://www.listbox.com/member/archive/247/2010/01/sort/time_rev/page/4/entry/11:270/20100118113841:EE029CD4-044F-11DF-868D-868DA52306B0/</a>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com1tag:blogger.com,1999:blog-17329179.post-81961265705534861492009-12-19T09:53:00.000-08:002009-12-19T10:42:10.518-08:00Why don't people use certificate-based access authentication?Most systems still only offer username/password authentication, and most people are still happy to use it, even though everyone knows (for example, through daily media headlines) that there are pervasive user access security problems with it.<br /><br />Why don't people use certificate-based access authentication?<br /><br />This question is important for email security and also in other areas, such as web site and blog access.<br /><br />We suggest that a proper answer requires thinking that has to be much more nuanced and sophisticated than just a discussion of usability versus security.<br /><br />Such thinking should come also from analyzing online and offline feedback, as we need to approach the question as it is seen -- from many sides.<br /><br />We have taken this approach in our paper, now updated, at <a href="http://email-security.net/papers/takefive.htm">http://email-security.net/papers/takefive.htm</a><br /><br />Please provide your comment. You can also <a href="http://email-security.net/papers/takefivecompact.htm">Read the Compact Version</a><br /><br />Thank you!<br />Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com10tag:blogger.com,1999:blog-17329179.post-5107739093624674082009-11-13T11:34:00.000-08:002009-11-14T19:19:57.584-08:00Let's "Take Five" In Internet SecurityWith everything that is happening (and not happening) in Internet security today, and all its complexity, it is perhaps useful to stop our busy day and take a little time out to start a conversation and question a couple things. <br /><br />The worst Internet security problem for users today is not email or even about email, however it deeply affects email security. We are talking about the security and usability of Internet user access control systems. This problem is well-known but we meekly accept it "as it is" everyday.<br /><br />But the paradigm may shift in five minutes. We find that, surprisingly, to tackle this problem we just need to take five minutes to go over five frequently asked questions. And that is our invitation to read the paper and provide your comment at <a href="http://email-security.net/papers/takefive.htm">http://email-security.net/papers/takefive.htm</a><br /><br />You can also <a href="http://email-security.net/papers/takefivecompact.htm">Read the Compact Version</a><br /><br />Thank you!<br />Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com8tag:blogger.com,1999:blog-17329179.post-38705158592101957342007-07-19T16:42:00.000-07:002007-07-20T11:10:27.462-07:00Discussion summary on improving SSHSSH (OpenSSH) is a workhorse for secure web administration and, as such, important for managing Internet servers including email. SSH is, usually, the only way to access remotely located servers. Of course, this makes SSH a valuable attack target.<br /><br />There are established practices for securing SSH. However, they currently rely substantially on additional systems, that also present their problems and add to the maintenance burden.<br /><br />After some discussion in the cryptography@metzdowd.com list and in private, I present below a summary of the dialogue and clarifications with a view to improve SSH. I've divided it in sections A-F, to make it easier to discuss.<br /><br /><span style="font-weight: bold;">A. THE SIX SSH ISSUES</span><br />The free test tool Nessus allows a remote SSH server to be quite easily discovered and explored by automated scans & attacks. The information obtained in SSH server scans can be typically divided (from the view point of a sys admin) in six issues (to be discussed later, this is just a list):<br /><br /><span style="font-style: italic;">1. no way to block scanning and attacks<br />2. does not learn from attacks<br />3. advertises itself and the server OS<br />4. allows empty authentication requests (eg, requests by invalid users)<br />5. sends host key fingerprint to invalid users (eg, with no username)<br />6. sends a reply to invalid users (eg, with no username)<br /></span><br />These issues, alone or in combination, can lead to multiple vulnerabilities, some with higher security significance than others (according to operational factors as well).<br /><br /><span style="font-weight: bold;">B. CRYPTO</span><br />Let's first state that not all the six issues are created by crypto use in SSH. And this leads us to the main question here:<br /><blockquote><span style="font-style: italic;">Could new crypto in SSH help mitigate or solve any of these six issues?<br /></span></blockquote>Currently, how are these issues addressed? By non-portable and potentially buggy methods (see below), with more software that has to be patched, updated, configured and tested, all at different times, with added security gaps that can be compromised.<br /><br />New crypto methods in SSH could do a better job by reducing the problems in the first line of defense at SSH itself and by avoiding security gaps. Non-crypto things such as PAM, tcpwrapper, port-knocking, grey-listing and others can always be added later anyway, if desired.<br /><br /><span style="font-weight: bold;">C. PASSWORDS</span><br />Some issues could be minimized by turning off password authentication, to use only private/public keys (PKs). SSH allows users to generate PKs without cost (no need for a CA). Users can keep their private keys in their drives and also in a USB dongle, for example. One can also <a href="http://www.joerg.cc/cgi-bin/wiki.pl?MySecurID-page">add second-channel authentication</a> (e.g., RSA SecurID)<span style="font-family: monospace;"> </span>to SSH with a PAM or a patch. <span style="font-style: italic;"><br /><br />However, turning off password authentication is not desired in many cases, even when PKs and second-channel authentication are used.</span><br /><br />Password authentication may be more secure against attacks on the client, to gather information before compromising the server. According to Peter Gutmann, a study of SSH attacks a few years ago showed that nearly two thirds of all SSH private keys were stored on disk with no password protection at all, so that simply being able to read a hard drive will get you access to any number of systems without having to trojan the SSH client or plant a keyboard logger as you'd need for an SSH password. So turning off password authentication could make things less secure, not more.<br /><br />Another reason for using passwords in SSH, and perhaps more compelling, is that you don't want to risk getting locked out just because your disk died, you forgot your laptop, SecurID or USB dongle, and the private key cannot be found (for security reasons!) in the back-up. Passwords can also easily be given over the phone from sys admin to sys admin, and are easily protected by threshold cryptography. And using a private/public key pair in SSH would anyway require a password for protecting the private key.<br /><br />Using good, high-entropy passwords turns out to be preferred by many sys admins because it is simpler, more portable, and works fine for security (provided the password is chosen with enough entropy). Sys admins already need to use many passwords for many servers and protocols, and have safe methods for off-line storage, redundant record keeping, and authorized record access (including a handwritten notebook, encrypted USB dongles, and a printed list), that are of course not kept near the computer.<br /><br /><span style="font-weight: bold;">D. PAM</span><br />Some of the issues can be solved with <a href="http://www.gsp.com/cgi-bin/man.cgi?section=8&topic=pam_ssh">SSH PAM</a>, such as <a href="http://www.hexten.net/wiki/index.php/Pam_abl">pam-abl</a>. However, <a href="http://lists.debian.org/debian-ssh/2006/12/msg00160.html">PAM libraries are usually buggy</a> (buggier than SSH), and might be unstable unless suitable PAM items were added by which the daemon can communicate the client IP address and port number, and maybe other things too.<br /><br /><span style="font-weight: bold;">E. SOLUTIONS</span><br />I believe that each one of the issues 1-6 can be met by a cryptographic protocol deployed in SSH. Some solutions are well-known. BTW, there were some similar issues in the SSL protocol.<br /><br /><span style="font-weight: bold;">F. ISSUE BY ISSUE REVIEW</span><br /><br /><span style="font-style: italic;">#1. no way to block scanning and attacks</span><br /><blockquote>Firewall port-knocking can help here. It is hard to guess and defeat; nothing shows up in scanning. However, it is not portable and is hard to debug and use.<br /><br />While there are also IDS that can turn off the SSH port if a scanning is detected, this can lead to DoS.<br /><br />Pam-abl can also help, but PAM is usually buggier and adds security gaps.<br /><br />Possible solution: SSH could easily incorporate some of the pam-abl code, enough to block an IP for some time T after N failed attempts, self-clearing after T.<br /></blockquote><span style="font-style: italic;">#2. does not learn from attacks</span><br /><blockquote>Possible solution: SSH can incorporate some of the pam-abl code for this purpose.<br /></blockquote><span style="font-style: italic;">#3. advertises itself and the server OS</span><br /><blockquote>"Need to know" is a valuable security policy.<br /><br />Advertising the OS may compromise a higher security function (the OS) in the case of a zero-day or unpatched exploit. SSH MUST NOT advertise the OS. It may compromise policies and apps in a lower (more critical) layer than SSH, not just the OS.<br /><br />Also, advertising SSH's innards in "already dead" authentication has no useful purpose that could not be more securely met.<br /><br />Possible solution: let the client tell its version to the server, for the server to decide if it is acceptable -- it is a server function anyway (eg, not to accept SSH v1).<br /><br />Additionally, it would also be useful for the client to send to the server a list of acceptable configurations, for the server to select, so that the server would no longer be groping in the dark as to what algorithm and parameters to use.<br /></blockquote><span style="font-style: italic;">#4. allows empty authentication requests (eg, requests by invalid users)<br />#5. sends host key fingerprint to invalid users (eg, with no username)<br />#6. sends a reply to invalid users (eg, with no username)</span><br /><blockquote>- Some claim that the SSH protocol requires key exchange to occur before the user identity is sent, as it protects the user identity from disclosure to an impersonator.<br /><br />However, if the user does not have a valid username, why and what should the server have to prove to that user? Invalid data deserves either "no response" (safer) or an error response. The SSH server should not do work or provide information to an user that is known already to be invalid. The SSH protocol needs improvement here.<br /><br />- Some also claim that if the SSH protocol would permit such differing responses for valid/invalid users, then attackers can use SSH as an oracle to determine username validity.<br /><br />However, SSH can prevent username enumeration by trial and error. With a simple keyed-hash, if this protection is desired, one would just enter the same key for the server and clients, so they can be all using the same transform. This is a simple mechanism that hackers cannot break by guessing. Note that SSH already uses a hash mechanism to hide domain names in the hosts file.<br /><br />- Some claim that the issues #4 and #5 are actually "features" that are used by some of the anonymous services, such as anonymous CVS.<br /><br />However, these "features" could be closed by default, and opened in the sshd config file if desired. This may also need some cryptographic protocol changes.<br /></blockquote>Because the reported issues could be improved by adding crypto, I am asking for more input first, before taking these issues to the OpenSSH development list. This should help the development effort. The OpenSSH development mailing list is at:<br /><a href="https://lists.mindrot.org/mailman/listinfo/openssh-unix-dev">https://lists.mindrot.org/mailman/listinfo/openssh-unix-dev</a><br /><br />Thanks for all the input! Comments are welcome.<br /><br />Cheers,<br />Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com8tag:blogger.com,1999:blog-17329179.post-1141085347632426522006-02-27T16:08:00.000-08:002006-02-27T16:09:10.783-08:00Re: NPR : E-Mail Encryption Rare in Everyday Use...<br /><br />I personally would prefer to sign every email I send. I'd also<br />prefer to encrypt all non-public messages. I am fully competent<br />in the use of the current technology, but it turns out to be not<br />practical to use.<br /><br />GregEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0tag:blogger.com,1999:blog-17329179.post-1141085154851581042006-02-27T16:02:00.000-08:002006-02-27T16:05:54.923-08:00Re: NPR : E-Mail Encryption Rare in Everyday UsePaul Hoffman wrote:<br /><blockquote type="cite">This is my original disagreement with Ed's message. It can be done, and when you do it it works, but it is too difficult for most people to bother with. I think we all agree on those three facts, just not on what to label the last one.<br /></blockquote><br />Actually, when I wrote "it does not actually work" I meant all three things:<br /><br />1. It can't be done as a user would like to do it; note also that even experts<br />do it incorrectly (it's just too many detail devils).<br /><br />2. When a user does it, the user does not really know if it was done right.<br /><br />3. It is too difficult for users to use and (worse) most users who use it<br />do it incorrectly.<br /><br />We have some choices. We can continue to say that it works and just wait<br />for users to get educated someday. Or, we can say that there is no x (x = market,<br />need, risk, point) -- and that's why no user bothers with it. Or, we can try<br />to understand what's it that users reject and work around it. My opinion I<br />already say upfront: users reject the whole model; it's not "natural" to<br />ask me for my envelope before you can send me a letter.<br /><br />(btw, name and mail address are not the envelope -- they are routing<br />information. My public-key is the envelope analogue when comparing postal mail<br />with secure email.)<br /><br />Cheers,<br />Ed GerckEd Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com1tag:blogger.com,1999:blog-17329179.post-1141084554318435412006-02-27T15:54:00.000-08:002006-02-27T15:55:54.390-08:00Re: NPR : E-Mail Encryption Rare in Everyday Use<pre wrap="">Phil Z doesn´t know how to do it himself, at least with PGP.<br />He told me that he doesn´t sign people´s keys who ask for it,<br />simply because it would pollute his keyring on his computer,<br />and he couldn´t work with a keyring with thousands of people<br />on it anymore.<br /><br />So PGP obviously has a usability and scalability problem.<br />So he only signs the keys of his friends because of that.<br />I wonder now, why he didn´t tried to solve that<br />usability/scalability problem himself yet, but gave up instead.<br /><br />Best regards,<br />Philipp Gühring<br /><br /><br /></pre>Ed Gerckhttp://www.blogger.com/profile/11500735527163002826noreply@blogger.com0