Saturday, April 2, 2005

How Physicists contributed to the Internet

Many people don't realize that their daily internet life has been affected more or less by works from physicists. Do you know who is the inventor of word-wide-web (HTTP, HTML and URL things)? Tim Berners-Lee, and he was a physicist working at CERN in europe. One of the gurus in TCP? Van Jacobson, who was working in high energy physics at Berkeley's Lawrence Laboratory.

I believe many physicists did contribute for the improvement and inventions of the current TCP/IP and other protocols currently in use.

Wednesday, March 23, 2005

Computer Reuse Through Linux

One day I found a laptop dumped in recycle bin at my office. It is Toshiba Tecra 8100 with Pentium3 450 MHz, 64-MB RAM and 12 MB harddisk. It had Windows NT in it. I took it home, reformatted it with my SuSE Linux 9.2. Well, the memory seemed not enough, so I went to ebay.com and found somebody was selling the 128 RAM SIMM100 for around 20 bucks. Not bad, I though. So I bought it and installed it on the laptop.

I got 192 MB now so I could run the GUI (I use KDE, but also added many GNOME libraries to run GNOME-based applications). I then downloaded the latest kernel available at that time (2.6.10). I also bought Cisco Linksys PCMCIA WLAN card (WPC54 SpeedBooster). Unfortunately, Linksys has not provided the driver for Linux yet, but luckily there is ndiswrapper downloadable somewhere. So I copied the driver for Windows to my Linux, run the ndiswrapper and ...voila, the wireless card worked. Well, still had problem here. Apparently, there was a conflict between ndiswrapper and ndiswrapper. I rebuilt the kernel and disabled sound drivers, but still sporadically the ndiswrapper did work very well (sometimes, the WLAN lost connection). For your info, I rebuilt them with specific processor criterias enabled, such as mcpu=pentium3 -msse and mfpmath=sse.

A few days I go, I gave a try to use kernel 2.6.11.5. I even rebuilt Krolltech's QT and many libraries. After many days of recompiling, I successfuly made the wireless work with sound drivers. I was one of the happiest days in my life making reuse the old laptop. I have been using the laptop for many of my daily activities, including browsing, reading emails and even a lightweight server. Yes, it is a server. Imagine if I use Windows for this purpose, I might have burned the laptop to the hell for its slowiness.

The laptop has SSH server, FTP, Telnet and many other services. I even also connect my external USB harddrive, thus I got additional 12 GB of space for storage. Not bad at all.

At work, I also partition my other laptop (it is IBM T30 with 512 MB RAM and 40 GB of total space). I parition 6 GB for Linux, and the rest for Windows 2000. You know what? I ended up using Linux for my work activities almost everyday. Linux is really cool, and I have learned a lot about many things because of opensource applications and tools available from the internet.

I really thank people outthere who have developed such great operating systems, applications and tools.

Thursday, March 3, 2005

Got Answer from one of the 'Hacker'

A few weeks ago I modified an article about SHA-1 on www.wikipedia.org by adding a link to another page telling a brief biography about one of the SHA-1 hackers, a chinese researcher name Xiaoyun Wang. After a few minutes, somebody removed the link and even the new page I added due to infringement of copyrighted materials.

I was suprised, but then I sent email to the researcher asking wether she objects my writing. I got a reply few days later saying that her team and the university (Shandong University of China) are going to create a new website dedicated to this security stuff. Well, she did not really answer my questions, but at least I got a response from an expert and Ph.D in security.

Let's wait and see how their website and papers will look like.

Rebuilding KDE made Easy!

After a few months not checking KDE website (www.kde.org), two days ago I revisited the site and found an interesting tool for KDE 3.3.2: Konstruct. The tool is easy to use and is designed to build (checking components and to download the missing ones, configure them, compile and link the whole libraries and component).

The only command I needed to execute is:

cd konstruct/meta/kde; make install

So easy to build now. If I recall, it was giving me hard time to recompile my KDE (it was 3.3) on my IBM Laptop T30. I had to download all *.bz2 files (plus qt-x11 libraries), extract them, reconfigure and compile one by one.

I am still having problem though when compile them on my 'free' Toshiba Laptop Tecra8100. Somehow, one component (Kppp) complains about 'regfree' and some other procedures altough I have double checked the Kpp*.cpp has "#include ". Anybody knows how to resolve it?

Friday, February 25, 2005

Crack in SHA-1 code 'stuns' Security Gurus

Three chinese researchers said on February 14, 2005 that they have compromised the SHA-1 hashing algorithm at the core of many of today's mainstream security products.

In the wake of the news, some cryptographers called for an accelerated transition to more robust algorithms and a fundamental rethinking of the underlying hashing techiques.

"We've lost our safety margin, and we are on the edge," said William Burr, who manages the security technology group at the National Institute of Standards and Technology (NIST).

"This will create big waves, in my opinion," said the celebrated cryptographer and co-inventor of SHA-1 (Shamir hashing Alg.), Adi Shamir. "This break of SHA-1 is stunning," concurred Ronald Rivers, a professor at MIT who co-developed the RSA with Shamir.

RSA is a public-key cryptosystem for both encryption and authentication; it was invented in 1977 by Ron Rivest, Adi Shamir, and Leonard Adleman [RSA78]. Details on the algorithm can be found in various places. RSA is combined with the SHA1 hashing function to sign a message in this signature suite. It must be infeasible for anyone to either find a message that hashes to a given value or to find two messages that hash to the same value. If either were feasible, an intruder could attach a false message onto Alice's signature. The hash functions SHA1 has been designed specifically to have the property that finding a match is infeasible, and is therefore considered suitable for use in this role.

One or more certificates may accompany a digital signature. A certificate is a signed document that binds the public key to the identity of a party. Its purpose is to prevent someone from impersonating someone else. If a certificate is present, the recipient (or a third party) can check that the public key belongs to a named party, assuming the certifier's public key is itself trusted. These certificates can be held in the Attribution Information section of the DSig 1.0 Signature Block Extension and thus passed along with the signature to aid in validating it. (See section Attribution Information section in the DSig 1.0 Specification.)

The signature section of the DSig 1.0 Signature Block Extension is defined in the DSig 1.0 Specification. For the RSA-SHA1 signature suite, the signature section has the following required and optional fields.

Who are these three chinese researchers? One of the member, Lisa Yin was a Ph.D student who studied under Ronald Rivest (RSA inventor) at MIT. Another one was responsible for cracking the earlier MD5 hashing algorithm (also developed by Rivest in 1991) which happened in August 2004.

To learn more about MD5, please visit http://en.wikipedia.org/wiki/MD5. For RSA: http://en.wikipedia.org/wiki/RSA, and for SHA-1: http://en.wikipedia.org/wiki/SHA-1

The open-source code version of the algorithm can be found in http://www.cr0.net:8040/code/crypto/sha1/. Samir et.al published their paper at ACM forum: The RSA Encryption Algorithm, R.L. Rivest, A. Shamir, L.M. Adleman, "A method of Obtaining Digital Signatures and Public-Key Cryptosystems", Communications of the ACM, v. 21, n. 2, Feb. 1978, pp 120-126.

Electronics, biology: twins under the skin

By Chappell Brown
EE Times
February 09, 2004 (10:33 AM EST)

Like the twin strands of a double helix, electronics and biotechnology are joining forces in a technological explosion that experts say will dwarf what is possible for either one of them alone.

Hints of this pairing can be seen in the economic recovery that's now taking hold. One peculiarity that hasn't grabbed many headlines is biotech's role in pulling Silicon Valley out of its three-year slump. A report last month from the nonprofit organization Joint Venture: Silicon Valley Network points up this fact, showing that venture funding in biotech startups rose from 7 percent in 2000 to 24 percent last year while investment in information technology startups fell from 10 percent to 4 percent over the same period. The immediate question is whether this is a temporary anomaly or the emergence of a major trend.

Certainly computers, biochips, robotics and data sharing over the Internet have been important tools in accelerating biological and medical research, and it should be no surprise that new application areas and markets would grow around them. The view from inside the engineering cubicle might be something like, "Yes, we have created a revolutionary technology that creates new markets-biomedicine is simply one area that benefits from advances in VLSI."

But a long-term perspective suggests a tighter linkage between electronics technology and molecular biology. Indeed, it could be argued that the second half of the 20th century forged not one but two digital revolutions, fueled by two fundamental breakthroughs: transistorized digital computers and the cracking of the genetic code. The latter advance showed that the genome was transmitted through the generations by means of digital storage in the DNA molecule.

In the following decades, both developments matured at an increasingly rapid pace. Digital circuits were inspired by crude models of the nervous system (see story, below). Although the models turned out to be wrong in many respects, technologists discovered that digital representation brings the advantages of simplicity, stability and an ability to control errors. Those same properties have made DNA the viable and stable core of living systems for billions of years.

But the nervous system is only one component of the body that is encoded in DNA, which somehow not only represents the information for building the basic components of cells, but also encodes the entire process of assembling highly complex multicellular machines. The growth process is an amazing feat of bootstrapping from the genetic code to functioning organisms. Essentially, an organism is a molecular digital computer that constructs itself as part of the execution of its code.

Leroy Hood, director of the Institute for Systems Biology (Seattle), believes that science aided by computers and VLSI technology will achieve major breakthroughs in reverse-engineering the cell's assembly processes. The fallout will be new circuit and computational paradigms along with nanoscale mechanisms for building highly compact molecular computing machines.

"There will be a convergence between information technology and biotechnology that will go both ways," said Hood. "We can use new computational tools to understand the biological computational complexities of cells, and when we understand the enormous integrative powers of gene regulatory networks we will have insights into fundamentally new approaches to digital computing and IT."

But cell machinery can also be enlisted in the kind of nanostructure work that is currently done manually with tools such as the atomic-force microscope. "The convergence of materials science and biotech is going to be great, and we will be able to learn from living organisms how they construct proteins that do marvelous things and self-assemble," Hood said. "There will be lessons about how to design living computer chips that can self-assemble and have enormous capacity."

Hood is credited with inventing the automated DNA-sequencing systems that were the first step in accelerating the decoding of the human genome. Accomplished two years ahead of schedule thanks to many enhancements to the process, including MEMS-based microfluidic chips, the achievement has stimulated efforts to take on the far more complex task of decoding protein functions.

Hood's institute, which was founded in 2000, is one example of a wave of similar organizations springing up across the United States. The idea is to engage a diverse group of specialists-mechanical and electronic engineers, computer scientists, chemists, molecular biologists-in the effort to decode the cellular-growth process. Stanford University's BIO-X Biosiences Initiative, for example, is dedicated to linking life sciences, physical sciences, medicine and engineering. The Department of Energy's Pacific Northwest National Laboratory has its Biomolecular Systems Initiative, Princeton University its Lewis-Sigler Institute for Integrative Genomics. Harvard Medical School now has a systems-biology department, and MIT has set up its Computational and Systems Biology Initiative (CSBi).

Proteins have remarkable chemical versatility and go far beyond other molecules as chemical catalysts, said Peter Sorger, director of MIT's CSBi. But applications of their properties will have to contend with a difficult cost differential between medical and industrial products.

"Using proteins as catalysts was the absolute beginning of the biotech industry. We know that proteins are the most extraordinary catalysts ever developed. The problem is that most of the chemical industry is a low-margin business and biology has historically been very expensive," Sorger explained.

While organic catalysts derived from oil are not as efficient, the low cost of producing them has kept proteins out of the field. "Most of the applications of proteins to new kinds of polymers, new plastics, biodegradable materials, etc. have all been limited by the fundamental economic problem that oil is so darn cheap," he said. "As a result, bioengineered materials are only used in very high-end, specialized applications."

However, Sorger believes that such bioengineered products will arrive, probably first in biomedical applications, which will then spawn low-end mass-market products. He used the example of Velcro, which was devised as an aid to heart surgery and later became a common material in a wide range of commercial goods. Sorger is looking forward to nanotechnology applications, the assembly of materials and circuits using biological processes, as the first direct applications of protein engineering outside of the biomedical field.

Sorger cited the work of MIT researcher Angela Belcher as an example of the technological spin-offs that will come from attempts to understand cellular processes. Working in the cross-disciplinary areas of inorganic chemistry, biochemistry, molecular biology and electrical engineering, Belcher has found ways to enlist cellular processes to assemble structures from inorganic molecules. By understanding how cells direct the assembly of their structural components, Belcher is finding ways to assemble artificial materials from inorganic nanoclusters that can function as displays, sensors or memory arrays. Another interdisciplinary group at MIT is putting together a library of biological components that engineers could used to build artificial organisms able to accomplish specific nanoscale tasks.

Underlying the excitement surrounding the merger of digital electronics systems and molecular digital organisms are the dramatic capabilities of lab-on-a-chip chemical-analysis systems, automated data extraction and supercomputer data processing. These technologies are part of what made it possible to sequence the entire human genome. A benchmark for the rapid progress promised by those tools may be the announcement by three biotech companies late last year of single chips containing the human DNA molecule in addressable format-the human genome on a chip. That might compare to the advent of of the CPU-on-a-chip, which catalyzed the VLSI revolution in the mid-1970s.

The barrier to moving this capability forward lies in the physical differences between DNA and the proteins it codes. Proteins are built from DNA sequences as linear sequences of amino acids that then spontaneously fold into complicated 3-D shapes. And the process becomes more complex as proteins begin to interact with one another. For example, there is a feedback loop in which proteins regulate the further expression of proteins by DNA. As a result, there are no parallel fluidic-array techniques to accelerate the analysis of protein families. "These technologies have a long way to go. I don't see any fundamental breakthroughs [in protein analysis] in the next few years, but in 10 years, who knows?" said Steven Wiley, director of the Biomolecular Systems Initiative at Pacific Northwest National Laboratory. "There are a lot of smart people out there working on this."

The fundamental challenge is the dynamic aspect of protein function. "DNA is static; once you sequence it, you have it," Wiley said. But proteins "are constantly interacting, so you have to run multiple experiments to observe all their functions and you end up with multiple terabytes of information. So, how are you going to manage and analyze all this information?"

But the excitement generated by recent successes with the genome is contagious. Plans are afoot to decode the "language" of proteins, making their functions widely available to engineers; anyone with a personal computer and a modem can access the human genome over the Internet; lab-on-a-chip technology continues to reduce the cost of bioexperimentation while ramping up throughput. And there is venture capital funding out there.