Future Trends in Microelectronics - 95 - Kluwer
Future Trends in Microelectronics - 95 - Kluwer
Future Trends in Microelectronics - 95 - Kluwer
Microelectronics
Reflections on the Road to Nanotechnology
edited by
Serge Luryi
Department of Electrical Engineering,
State University of New York,
Stony Brook, NY, U.S.A.
Jimmy Xu
Department of Electrical & Computer Engineering,
University of Toronto,
Toronto, Ontario, Canada
and
Alex Zaslavsky
Division of Engineering,
Brown University,
Providence, Rl, U.S.A.
m
Kluwer Academic Publishers
no
Dordrecht / Boston / London
ro
Proceedings of the NATO Advanced Research Workshop on
Future Trends in Microelectronics: Reflections on the Road to Nanotechnology
lie de Bendor, France
July 17-21,1995
A C.I.P. Catalogue record for this book is available from the Library of Congress
ISBN 0-7923-4169-4
Form Approved
REPORT DOCUMENTATION PAGE OMB NO. 0704-0188
Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources,
gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comment regarding this burden estimates or any other aspect of this
collection of information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for information Operations and Reports, 1215 Jefferson
Davis Highway, Suite 1204, Arlington, VA 22202-4302, and to the Office of Management and Budget, Paperwork Reduction Project (0704-0188), Washington, DC 20503.
1. AGENCY USE ONLY (Leave blank) 2. REPORT DATE 3. REPORT TYPE AND DATES COVERED
Nov 96 Final
4. TITLE AND SUBTITLE 5. FUNDING NUMBERS
17. SECURITY CLASSIFICATION 18. SECURITY CLASSIFICATION 19. SECURITY CLASSIFICATION 20. LIMITATION OF ABSTRACT
OR REPORT OF THIS PAGE OF ABSTRACT
UNCLASSIFIED UNCLASSIFIED UNCLASSIFIED UL
NSN 7540-01-280-5500 Standard Form 298 (Rev. 2-89)
Prescribed by ANSI Std. 239-18
298-102
GENERAL INSTRUCTIONS FOR COMPLETING SF 298
The Report Documentation Page (RDP) is used in announcing and cataloging reports. It is important
that this information be consistent with the rest of the report, particularly the cover and title page,
instructions for filling in each block of the form follow. It is important to stay within the lines to meet
optical scanning requirements.
Block 1. Agency Use Only (Leave blank) Block 12a. Distribution/Availability Statement.
Denotes public availability or limitations. Cite any
Block 2. Report Date. Full publication date availability to the public. Enter additional
including day, month, and year, if available (e.g. limitations or special markings in all capitals (e.g.
1 Jan 88). Must cite at least year. NORFORN, REL, ITAR).
Block 3. Type of Report and Dates Covered. DOD See DoDD 4230.25, "Distribution
State whether report is interim, final, etc. If Statements on Technical
applicable, enter inclusive report dates (e.g. Documents."
10 Jun 87 - 30 Jun 88). DOE See authorities.
NASA See Handbook NHB 2200.2.
Block 4. Title and Subtitle. A title is taken from NTIS Leave blank.
the part of the report that provides the most
meaningful and complete information. When a
report is prepared in more than one volume, Block 12b. Distribution Code.
repeat the primary title, add volume number, and
include subtitle for the specific volume. On DOD Leave blank
classified documents enter the title classification DOE Enter DOE distribution categories
in parentheses. from the Standard Distribution for
Unclassified Scientific and Technical
Block 5. Funding Numbers. To include contract Reports
and grant numbers; may include program NASA Leave blank.
element number(s), project number(s), task NTIS Leave blank.
number(s), and work unit number(s). Use the
following labels:
Block 13. Abstract. Include a brief (Maximum
C - Contract PR - Project 200 words) factual summary of the most
G - Grant TA - Task significant information contained in the report.
PE - Program WU - Work Unit
Element Accession No.
Block 14. Subject Terms. Keywords or phrases
Block 6. Author(s). Name(s) of person(s) identifying major subjects in the report.
responsible for writing the report, performing
the research, or credited with the content of the
report. If editor or compiler, this should follow Block 15. Number of Pages. Enter the total
the name(s). number of pages.
Block 7. Performing Organization Name(s) and
Address(es). Self-explanatory. Block 16. Price Code. Enter appropriate price
code (NTIS only).
Block 8. Performing Organization Report
Number. Enter the unique alphanumeric report Block 17.-19. Security Classifications. Self-
number(s) assigned by the organization explanatory. Enter U.S. Security Classification in
performing the report. accordance with U.S. Security Regulations (i.e.,
UNCLASSIFIED). If form contains classified
Block 9. Sponsoring/Monitoring Agency Name(s) information, stamp classification on the top and
and Address(es). Self-explanatory. bottom of the page.
Block 10. Sponsoring/Monitoring Agency
Report Number. (If known) Block 20. Limitation of Abstract. This block must
be completed to assign a limitation to the
Block 11. Supplementary Notes. Enter abstract. Enter either UL (unlimited) or SAR (same
information not included elsewhere such as; as report). An entry in this block is necessary if
prepared in cooperation with...; Trans, of...; To be the abstract is to be limited. If blank, the abstract
published in.... When a report is revised, include is assumed to be unlimited.
a statement whether the new report supersedes
or supplements the older report.
Standard Form 298 Back (Rev. 2-89)
Future Trends in Microelectronics
NATO ASI Series
Advanced Science Institutes Series
A Series presenting the results of activities sponsored by the NATO Science Committee,
which aims at the dissemination of advanced scientific and technological knowledge,
with a view to strengthening links between scientific communities.
The Series is published by an international board of publishers in conjunction with the NATO
Scientific Affairs Division
PARTNERSHIP SUB-SERIES
1. Disarmament Technologies Kluwer Academic Publishers
2. Environment Springer-Verlag / Kluwer Academic Publishers
3. High Technology Kluwer Academic Publishers
4. Science and Technology Policy Kluwer Academic Publishers
5. Computer Networking Kluwer Academic Publishers
NATO-PCO-DATA BASE
The electronic index to the NATO ASI Series provides full bibliographical references (with keywords
and/or abstracts) to more than 50000 contributions from international scientists published in all
sections of the NATO ASI Series.
Access to the NATO-PCO-DATA BASE is possible in two ways:
- via online FILE 128 (NATO-PCO-DATA BASE) hosted by ESRIN,
Via Galileo Galilei, I-00044 Frascati, Italy.
- via CD-ROM "NATO-PCO-DATA BASE" with user-friendly retrieval software in English, French
and German (© WTV GmbH and DATAWARE Technologies Inc. 1989).
The CD-ROM can be ordered through any member of the Board of Publishers or through NATO-
PCO, Overijse, Belgium.
cH
LHP
Series E: Applied Sciences - Vol. 323
This book contains the proceedings of a NATO Advanced Research Workshop held within the programme
of activities of the NATO Special Programme on Nanoscale Science as part of the activities of the NATO
Science Committee.
Other books previously published as a result of the activities of the Special Programme are:
NASTASI, M., PARKING, D.M. and GLEITER, H. (eds.), Mechanical Properties and Deformation Behavior of
Materials Having Ultra-Fine Microstructures. (E233) 1993 ISBN 0-7923-2195-2
VU THIEN BINH, GARCIA, N. and DRANSFELD, K. (eds.), Nanosources and Manipulation of Atoms under
High Fields and Temperatures: Applications. (E235) 1993 ISBN 0-7923-2266-5
LEBURTON, J.-P., PASCUAL, J. and SOTOMAYOR TORRES, C. (eds.), Phonons in Semiconductor Nanostruc-
tures. (E236) 1993 ISBN 0-7923-2277-0
AVOURIS, P. (ed.), Atomic and Nanometer-Scale Modification of Materials: Fundamentals and Applica-
tions. (E239) 1993 ISBN 0-7923-2334-3
BLÖCHL, P. E., JOACHIM, C. and FISHER, A. J. (eds.), Computations for the Nano-Scale. (E240) 1993
ISBN 0-7923-2360-2
POHL, D. W. and COURJON, D. (eds.), Near Field Optics. (E242) 1993 ISBN 0-7923-2394-7
SALEMINK, H. W. M. and PASHLEY, M. D. (eds.), Semiconductor Interfaces at the Sub-Nanometer Scale.
(E243) 1993 ISBN 0-7923-2397-1
BENSAHEL, D. C, CANHAM, L. T. and OSSICINI, S. (eds.), Optical Properties of Low Dimensional Silicon
Structures. (E244) 1993 ISBN 0-7923-2446-3
HERNANDO, A. (ed.), Nanomagnetism (E247) 1993. ISBN 0-7923-2485-4
LOCKWOOD, DJ. and PINCZUK, A. (eds.), Optical Phenomena in Semiconductor Structures of Reduced
Dimensions (E248) 1993. ISBN 0-7923-2512-5
GENTILI, M., GIOVANNELLA, C. and SELCI, S. (eds.), Nanolithography: A Borderland Between STM, EB,
IB, and X-Ray Lithographies (E264) 1994. ISBN 0-7923-2794-2
GÜNTHERODT, H.-J., ANSELMETTI, D. and MEYER, E. (eds.), Forces in Scanning Probe Methods (E286)
1995. ISBN 0-7923-3406-X
GEWIRTH, A.A. and SIEGENTHALER, H. (eds.), Nanoscale Probes of the Solid/Liquid Interface (E288)
1995. ISBN 0-7923-3454-X
CERDEIRA, H.A., KRAMER, B. and SCHÖN, G. (eds.), Quantum Dynamics of Submicron Structures (E291)
1995. ISBN 0-7923-3469-8
WELLAND, M.E. and GIMZEWSKI, J.K. (eds.), Ultimate Limits of Fabrication and Measurement (E292)
1995. ISBN 0-7923-3504-X
EBERL, K, PETROFF, P.M. and DEMEESTER, P. (eds.), Low Dimensional Structures Prepared by Epitaxial
Growth orRegrowth on Patterned Substrates (E298) 1995. ISBN 0-7923-3679-8
MARTI, O. and MÖLLER, R. (eds.), Photons and Local Probes (E300) 1995. ISBN 0-7923-3709-3
GÜNTHER, L. and BARBARA, B. (eds.), Quantum Tunneling of Magnetization - QTM '94 (E301) 1995.
ISBN 0-7923-3775-1
PERSSON, B.N.J. and TOSATTI, E. (eds.), Physics of Sliding Friction (E311) 1996. ISBN 0-7923-3935-5
MARTIN, T.P. (ed.), Large Clusters of Atoms and Molecules (E313) 1996. ISBN 0-7923-3937-1
DUCLOY, M. and BLOCH, D. (eds.), Quantum Optics of Confined Systems (E314). 1996.
ISBN 0-7923-3974-6
ANDREONI, W. (ed.), The Chemical Physics of Fullereness 10 (and 5) Years Later. The Far-Reaching
Impact of the Discovery ofC^ (E316). 1996. ISBN 0-7923-4000-0
NIETO-VESPERINAS, M. and GARCIA, N. (Eds.): Optics at the Nanometer Scale: Imaging and Storing with
Photonic Near Fields (E319). 1996. ISBN 0-7923-4020-5
LURYI, S., XU, J. and ZASLAVSKY, A. (Eds.): Future Trends in Microelectronics: Reflections on the Road to
Nanotechnology (E323). 1996. ISBN 0-7923-4169-4
RARITY, J. and WEISBUCH, C. (Eds.): Microcavities and Photonic Bandgaps: Physics and Applications
(E324). 1996. ISBN 0-7923-4170-8
CONTENTS
CONTRIBUTORS 413
INDEX 417
Preface
Ever since the invention of the transistor and especially after the advent of
integrated circuits, semiconductor devices have kept expanding their role in our life. For
better or worse, our civilization is destined to be based on semiconductors.
Identifying the scenarios for the future evolution of microelectronics is the key
to constructive action today.
Perhaps this can be dismissed as "fortune telling" or, at best, viewed as a high
risk undertaking. Indeed, prediction is hard to make, especially when it is about the
future. And, too often did forecasts by well-informed and authoritative sources prove
wildly wrong. A few examples can be readily found in the field of computers - the most
valued "customer" of microelectronics:
XI
Xll
"I have traveled the length and breadth of this country and talked
with the best people, and I can assure you that data processing is a fad that
won't last out the year."
- An editor for Prentice Hall, 1957.
However, advocates for new approaches and optimists of departure from the
existing path of established technologies usually fare no better. Critics of adventurous
new approaches, though not often cited, are frequently in the right. Even worse, we
have all seen superior technologies fail for reasons entirely unrelated to technical
merits...
Still, as we shed our illusions, we can not afford actions without vision.
What is needed is critical assessment of where we are, what lies ahead, where
new opportunities and/or alternative paths might be and what the limiting factors are...
• What kind of research does the silicon industry need to continue its
expansion? What are the anticipated trends in lithography? Modeling?
Materials? Can we expect a "display revolution"? Will wide-area
electronics be integrated with VLSI?
• To what extent can we trade high speed for low power? Is adiabatic
computing in the cards?
• Is there a need for (possibility of) integrating compound semiconductor
IC's into Si VLSI? What are the merits and prospects of hybrid schemes,
such as heteroepitaxy and packaging? What are the most attractive system
applications of optoelectronic hybrids?
This book is a result of this exercise. It is a reflection of the issues and views
debated at the workshop and a summary of the technical assessments and results
presented.
For all those who came together to share ideas, and for all the prospective
readers, we hope that this publication will serve as a useful reference and a springboard
for new ideas.
HERBERT KROEMER
ECE Department, University of California
Santa Barbara, CA 93106, USA
1. Introduction
At the 1974 International Electron Devices Meeting (IEDM), Marty Lepselter gave an
invited talk under the title "Integrated circuits - The New Steel" [1]. His message was
that the emerging integrated circuit technology was likely to play the same central role
in the industrial revolution of the late-20th century that steel played in the great indus-
trial revolution of the early-19th century.
I have always found this an extraordinarily apt analogy, and it has long been my
conviction that this analogy can be extended to an important general analogy between
structural metallurgy in general and electronic metallurgy in general [2].
In structural metallurgy, steel has, for some two centuries, been the dominant
structural metal—and is likely to remain so for the foreseeable future. Without steel, we
would have no modern industrial society. But we also would not have such a society if
we relied on steel alone. Modern society depends vitally on the diversity contributed by
such additional materials as aluminum, magnesium, titanium, etc. While we continue to
build automobiles and ships and other "heavy goods" (mostly) from steel, if steel were
the only structural metal available, we would still build airplanes from wood, and we
would not build spacecraft (and communication satellites) at all.
Similarly, in electronic metallurgy, Silicon is, without ant doubt, the dominant
electronic material—and is likely to remain so. But a mature electronic technology, too,
calls for a great diversity, more than can be provided by Si technology alone. Enter
other materials, such as GaAs and beyond.
Structural metallurgists divide metallurgy into ferrous and nonferrous metallurgy.
The analogy to Si and compound semiconductor technology is obvious. In a very real
The use of the new technology to obtain merely better quantitative improvements in
applications for which a technology already exists always has been—and will continue
to be—a secondary consequence of the success of the new technology in new appli-
cations, usually as a result of cost reductions brought about by the extensive use of the
new technology in the new applications.
The pattern of new science creating new devices that create their own applications
is likely to continue well into the next century.
2.3. EXAMPLES
2.4. LESSONS
If it is indeed true that the principal applications of any sufficiently new and innovative
technology will be applications created by that new technology, then this has the far-
reaching consequence that all of us must take a long-term look when judging the
potential of any new technology:
Too many attempts to look at the future of semiconductor judge new device concepts
by whether they can be mass-produced at the huge volumes and low cost that are
characteristic of Si integrated circuit technology. This is of course appropriate for
concepts that are indeed intended to find their application in the same market as Si
integrated circuits, where it is indeed extraordinarily difficult to compete with the
existing technology.
But remember that the applications of new concepts are more likely to be applica-
tions that get generated by the new concepts than pre-existing applications, and here
the economics is an altogether different one. What matters for the economic viability of
the new technology is simply whether the added value of the new application can
support the R&D cost and the manufacturing cost of that technology. If a new technol-
ogy has enough of that crucial economic leverage I referred to earlier in the context of
HEMT's, it may be economically viable even at a low manufacturing volume and a
high attendant cost per device. For example, if a new but expensive-to-make $1000
device makes possible a new $20,000 instrument that simply cannot be built without
that device, and if there is enough demand for the enhanced capability of that instru-
ment to permit a recovery of the cost of making each device, then the technology for
making the device becomes self-supporting, and has a chance of surviving—never
mind that the increase in cost over, say, silicon technology is huge: The latter cannot do
the job. Recent history abounds with examples of such high-leverage devices, and one
of my predictions is that we will see much more of this, especially in the instrumen-
tation and sensor field, and that high-leverage applications in these fields will be
amongst the driving engines of device technology for the next century.
The number of such devices for any single such application, and even their asso-
ciated money value, may be minuscule compared to the number and money volume of
Si IC's, but this does not in any way diminish the attractiveness of the devices to those
working on them: Working on high-leverage special-purpose devices may, in fact, be
an attractive career path for a young scientist or engineer. Moreover, it is an excellent
way for universities to prepare future scientists and engineers for the technologies of
the future. Nor are such high-leverage activities negligible from the point of view of the
economics of an entire nation: While each individual example might indeed have a
negligible impact on that economics, the cumulative effect of the very large number of
such activities can be huge.
The idea that the principal applications of new technology will be applications created
by that technology, calls for an assessment of the role of open-ended research, not tied
to a specific application.
What I call open-ended research is often referred to simply as long-term research,
but long-term research need not be open-ended, as some of the discussion at this
workshop on the development of CMOS technology past the year 2010 demonstrates.
Some call it curiosity-driven, as opposed to being applications-driven, but this, too,
does not hit the mark: While the motivation of the individual researcher may very well
be pure curiosity, those of society at large, which supports this activity, are not: Ulti-
mately, society does expect a payoff even from open-ended research, be it direct or
indirect. Society is simply willing to leave it open what that payoff might be, based on
the experience that there always has been such a payoff, not necessarily on every
project, but certainly collectively. This specifically includes the recognition that the
payoff has often been indirect, through its impact on subsequent research one or more
research generations down the road.
One of the developments of the last decade that has deeply influenced and even
shocked all of us—and continues to do so—is the retreat of industry from this open-
ended research. I shall not analyze here the reasons why this happened, nor bemoan it,
but simply take it as a given that is likely to remain with us, and look at some of the
consequences, and specifically on the impact of this development on the universities: It
may very well turn out that the only places where open-ended research can be conduct-
ed in the future on a significant scale will be the universities. Let us turn to this issue.
3.2. THE ROLE OF THE UNIVERSITIES
I believe it is essential that those of us who are engaged in this kind of work tale a more
active role in bringing this last point to the attention of everybody else involved—so
they don't hear on;y the other side.
I will say relatively little about that part of my assertion that implies that society
needs a healthy open-ended research activity, if not in industry, then somewhere else.
Superficially, it appears that society's own recognition of this need has not changed.
For example, we all hear such buzzwords as industrial competitiveness, presumably
acknowledging a recognition of this need. But much of what I hear and see seems to be
more lip service than evidence of positive action; in fact, there is plenty of talk that
actually conflicts with a true recognition of that need. For example, the loud clamor for
more "relevance" even in non-industrial research can be safely translated into a clamor
for less open-ended research—and sometimes for less research of any kind.
Let me concentrate on the other part, that such research is an essential ingredient of
the educational mission of the university.
In some circles, the research done by the universities is not even viewed as meeting
an essential need of society, much less as an ingredient in education, but as a luxury
embarked on by an elite fraction (presumably a derogatory term) of faculty at those
universities that we call research universities—themselves only a fraction of all univer-
sities. Being a member of that elite (and not at all ashamed of it), I will be the first to
admit that I am gratified to be amongst those who are able to embark on this supposed
luxury. But is it really a luxury?
Furthermore, even though the total number of Ph.D.'s is only a fraction of the total
number of university graduates, even at most Ph.D.-granting research universities, this
fraction has a huge leverage, not only in industry, but also on all levels of education.
The Ph.D. as Future Teacher. The overwhelming majority of the teaching faculty in
science and technology at all universities, including at the undergraduate level, are
themselves products of Ph.D. programs. I stated earlier that the primary product of
universities is people; on the Ph.D.-level, one of the most important groups of those
people are those whose mission is as future educators at all university levels.
It is absolutely essential that those individuals must themselves have an education
that is up to the state of the art in their field. But just as important, they must be pro-
vided with the intellectual resources that permit them to stay at this forefront through-
out much of their careers, rather than having become obsolete carriers of past know-
ledge before the midpoint of that career. A research Ph.D. provides this obsolescence
reserve by providing the student with a knowledge the fundamentals that will be part of
the foundations of future developments as well. A Ph.D. research project does that more
thoroughly than mere coursework ever can, and it provides an opportunity to develop
the skills of actually applying those fundamentals in a real context. It doesn't really
matter much what the particular set of fundamentals is that is drawn upon in a given
research dissertation: It is the acquisition of a methodology that matters: The method is
the message!
10
The Ph.D. in Industry. All the above applies just as well to a career in industry, but
there is more. Going beyond research per se, open-ended research provides a "live edu-
cation" in what I like to call the management of uncertainty, the need to make rational
decisions in the face of very incomplete knowledge. This is a skill important to any
future leader in science and technology, including the teacher at a university, but it
becomes indispensable in industry, where a scientist or engineer often has to make
decisions about major commitments to future technology, in the face of precisely the
kind of very incomplete knowledge that characterizes open-ended research. The experi-
ence provided by the latter is likely to lead to more rational decisions than simply being
conservative (missing valuable opportunities), or following a "gut feeling," or follow-
ing the bandwagon "consensus" of equally-uninformed others. An additional merit of
an education involving open-ended research is that it tends to create people with a
personal stake in innovation, rather than in the maintenance of the status quo. Now we
all pay a lot of lip service to the need for innovation, often under such national banners
as industrial competitiveness, but much of it is again just lip service. I know only one
way to "institutionalize" innovation, and that is by bringing in young people and
relying on that most reliable of all human motivations, self-interest: They have nothing
to gain by maintaining the status quo, and everything by being the leaders of
innovation.
Properly understood and managed, the retreat of industry from open-ended research can
lead to a partnership in which both parties understand and serve their complementary
roles in society as partners rather than adversaries. But this partnership is not to come
about without an understanding of each other. There are problems on both sides.
The most serious threat coming from industry is the wide-spread clamor for "more
relevance" in university education. The reason for this is of course the desire of many
industrial managers to hire university graduates who make a positive contribution to the
"bottom line" from day-1, without requiring any further in-house training—let the long
term be damned! On the bachelor's level, economic realities may make this, to some
extent, inevitable (if deplorable). But it has very little, if any, validity on the Ph.D.
level, where it is nothing other than a thinly disguised call that universities abandon the
long-term research needs of society just as industry is abandoning them—and short-
changing their students along with society at large in the process.
We must resist these calls—but we must also communicate to society why! In the
last analysis, our obligations as educators are to our students and to society, not to
short-term-driven accountants at corporate headquarters, who—as recent years have
amply demonstrated—are increasingly viewing even the skills of their R&D staff as
11
short-term commodities rather than as valuable long-term investments. If career
changes are to be an essential ingredient in the future of industrial Ph.D.'s, we have
even more obligation to give them the broad long-term education that enables them to
make such changes. But that does not mean sacrificing a research education—to the
contrary.
Having said the above in defense of our research/teaching mission, I turn to the list of
things where we ourselves must do a better job as part of the bargain.
Probably the first need is that the university needs to communicate its role better to
society at large—much better. Because otherwise, society, and the political decision
makers whose decisions presumably reflect society's will, will only hear the other side,
which tends to be rather vocal.
Next in line is a need to recognized the changed role of the research Ph.D. itself.
This has several aspects.
We must stop pretending that we are educating our Ph.D.'s students for "pure"
research careers, academic or otherwise. Only a minority of them will enter such
careers; most will enter teaching careers or applied R&D careers (with more 'D' than
'R'), and we must make this clear to ourselves and our students:
A closely related need is to abandon a value system that rates "pure" research
somehow higher than applied research. What industry demands,—and justly so!—is
that we provide our students with a motivation towards the kind of shorter-term and
more applied work that is the very nature of industrial R&D, rather than creating people
who do that sort of thing only reluctantly. This is going to be one of the hardest changes
to make, because it is ultimately a cultural rather than scientific change, but it is
necessary all the same to make applied research at least as respectable as "pure"
research. This attitude seems to be highly developed in Japan, which goes a long way
towards explaining Japan's industrial success. A change in social values amongst the
rest of us might, in turn, go a long way helping us retaining our global competitiveness.
Finally, we must recognize that our Ph.D.'s will change their work many times in
the course of their careers, and that their education must enable them to handle such
changes. This calls for more breadth than is contained in a dissertation in which a
student spends too many years studying a single (usually rather narrow) topic in
"infinite-depth," often far beyond the depth the topic deserves. The Ph.D. should be a
12
certificate that says, in effect: "This individual has proven that he/she is capable to
perform independent high-quality engineering or scientific work, has adaptability to a
wide range of needs, and the ability to make, within a broad strategic context, the
decisions about how to conduct that work." This is far more useful than, say: "This
individual has spent over four years studying the low-temperature optical absorption of
sowhatnium, has honed the technique involved to perfection, and knows more about
this specific topic than anyone else in the world."
4. References
12 Statements
ARMIN W. WIEDER
Siemens Corporate Research and Development
Microelectronics
Otto-Hahn-Ring 6
81730 Munich, Germany
Abstract
The impressive progress in microelectronics in the last two decades has generated
enormous computational power and huge storage capacity at ever decreasing cost per
function. To carry on progress requires overcoming enormous challenges and likewise
will take advantage of great advances and opportunities. This translates into physics
overcoming all kinds of technical and technological limits. It requires technologists to
cleverly increase productivity and creatively define new products. Furthermore, it
requires overcoming the economic hurdles of exploding R&D costs and increasing
manufacturing investment. New opportunities, e.g. in the field of "system on chip",
will creatively integrate logic, memory and other functions on the same chip. New
schemes for cost-effective R&D and manufacturing will impose tremendous challenges,
especially on R&D personnel. These needs will require education, skills and creativity,
with much broader levels of expertise and knowledge ("research to production") in the
increasingly overlapping fields of devices, processes, circuits, chip architectures, product
definitions and manufacturing cleverness. The complex situation is condensed into 12
major statements.
1. Technology Perspectives
It took about 10 years after the invention of the transistor in 1948 to invent planar
technology and the integrated circuit: in 1959 the evolution of bipolar technology
started. It took another 10 years to overcome the interface problem of MOS devices, so
that the evolution of MOS circuits started in 1969. Since then every 3 years a new
generation of MOS technology emerges with 4 times the transistor count (complexity)
of the previous generation.
Today the leading semiconductor manufacturing companies are producing 4 Mbit
and 16 Mbit DRAM's and logic IC's with critical dimensions of 0.7 |im and 0.5 (Xm
respectively. Development in these companies is concentrating on 64 Mbit and 256
Mbit DRAM structures and the corresponding logic, with the 64 Mbit DRAM currently
entering the pre-production phase. Fully functional samples of the 256 Mbit DRAM's
13
5. Luryi et al. (eds.), Future Trends in Microelectronics, 13-21.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
14
Further progress sees no "show stoppers" for the next 15 years (1)
fT"
Lowe Air(b—nl—d)
Dielectrics
•BN«
Quantum Leap by New Materials ■Ö Papers <L
Electrode ••'HUvft (IMMWI ;id«)
Materials r\ IrOyQiwiwioxid«) ?
Fern
c Electrics
2
'S _ PLZT7ljead,-Unt«n.
Si y
Highe
Dielectrics
raffi • -.7..^.».—Bream»».
tio2 —- Titan*«)
?
(DRAM) B.$I.X^wti5tent«m.rJt»n««»)
- ■STtOtorniwi IT*»»«»)
03 TH4- --?•
c Metals
Cu/TOf :&r.:
wn-
M Suicides
.2
O
s=w
o €e8fe
2 ^RSh
„„„■MeSir-
P Si n§®£
ar±;
■.••■SiÖs
^-
ft
Figure 1.
15
The evolution of bipolar technology has resulted in integrated circuits with gate
delays below 20 ps, complexities of 105 components per chip and communication IC's
with data rates up to 40 Gbit/s and operating frequencies up to 30 Ghz. Further progress
will not encounter fundamental technical roadblocks, but bipolar technology is
challenged by the competing mainstream CMOS technology. Performance migration
and cost advantages of CMOS force bipolar more and more into analog and high
frequency and likewise force GaAs (and other III-V's) into super performance and
optoelectronic applications (Fig. 2).
CMOS technology wins in complexity, power dissipation and system speed (3)
|"H?MOS \^ ^i r \
s
/ I Inverter \ / / \ \
QED's have the potential to carry on microelectronics beyond year 2020 (4)
2. Economic Perspectives
Smart (accelerated) fabrication & development become the # 1 economic challenge (6)
Last but not least, decreased time to market and time to volume production require
concurrent engineering for development (whereby product development and process
development are completed simultaneously) as well as for concurrent manufacturing
design, ranging from equipment development to process development, product
development, facility design, and manufacturing operations.
The key performance yardsticks of modern microelectronics are 16 Mbit DRAM's with
40 ns access time; 300 k, 100 ps gate arrays; 0.5 (J,m CMOS with 107 components per
chip; and so on. This leads to figures of merit like computational power of about 200
MIPS. Future requirements of speech recognition and speech processing (e.g.
translation) with a vocabulary of 104 words in real time, however, will need about 10$
18
MIPS. Even more computational power will be needed for high quality (HDTV) real
time image/video recognition and processing: 106 MIPS. This also holds for
"intelligent" robots, expert systems, future man-machine-interface solutions, and other
sophisticated applications in information technology. All of these applications need an
increase of performance of 3 to 4 orders of magnitude.
This cannot be achieved by technological improvements alone. Novel architectures
are needed. All of them, however, need a massive increase in components per chip.
Therefore, by virtue of its superiority in yield, low power dissipation, design ease,
noise immunity, etc., CMOS will be the winning technology. Follow-up technologies
will have to be better in all aspects and not in just some of them.
The possibility of having on-chip memory and logic will lead to rationalized board
solutions of current products and also to novel systems not feasible today with
"quantum leaps" in performance. The latter will be realized by massive parallelism in
space (complexity) and time (pipelining). Neural networks, for instance, look best for
handling the complex problems of real-time man-machine interfaces. The trend to on-
19
chip memory and logic functions will cause microelectronics to go on integrating
forward into systems with increasing system knowledge and value added in silicon. This
will enforce much more intensive cooperation of system and semiconductor
manufacturers than in the past, as illustrated in Fig. 4 where future devices are classified
by system content rather than, say, minimum feature size in |j,m.
▲ RISC
• CISC
CO
Q.
1*
DEC alpha
PowerPC 620 ^ ^UitraSparc
0 w PentiurrK^ 4 \^k <3Micro2000
o
Q.
E
o *i386
a 10° ■'*.#■■'
Q.
=2.
3 MIPS
Under these conditions systems on-chip with "quantum leaps" in performance will
be manufactured. The dramatic progress can best be demonstrated by the prognosis that
in the year 2005 we will have the computational power of a CRAY-2 machine on a
single chip (Fig. 5)!
The computer-aided design and manufacturing (CAx) systems have to overcome
major challenges, too. On the one hand, deep submicron parasitics require CAx-systems
to deeply dig into physics, and on the other hand, the increased level of complexity
(Mega to Giga) requires solutions at higher abstraction levels. Furthermore, the aid of
computers and simulation tools is needed for questions ranging from physical, to
electrical, to economic: e.g. equipment, process, device, circuit and logic simulation,
accelerated yield learning, cost simulation, yield prediction, and so on all the way to
virtual factory and management tools. Here intensive cooperation between the
technology designer, the system architect, the system designer, the manufacturing
specialist, the manager, the business administrator, and the controller are mandatory in
order to provide relevant and cost-effective CAx-systems: ready for use simultaneously
with the corresponding microelectronic technology and the manufacturing facility.
The major changes in computer industry from 1980 to 1995 which led to the
decentralization of computers were primarily caused by the cost reduction of
computational power due to the progress of microelectronics. Today it appears that
further cost reduction and miniaturization of communication technologies, even greater
decentralization and networking of "computer intelligence" (PC's/workstations) and, last
but not least, the further reduction of cost of computational power will lead to the
redefinition of a new powerful industry emerging from communication, data and
consumer industry. TV's will become "intelligent" and gain communication
capabilities, the phone will be upgraded by "intelligent" displays and portability,
whereas the computer will get additional features like communications and portability.
Key competencies for this industry will be batteries, displays and low power IC
techniques like uP's, DSP's, uC's, DRAM's, E2PROM's, telecommunications IC's, ..
Microelectronics will grow twice as fast as the world electronics market (11)
This, however, implies exponential growth for another 15 years: an exponential growth
of "technical intelligence" (complexity, computational power, transfer rates, speed, ..).
This gigantic increase in "technical intelligence" could offer the means for solving the
serious problems in various areas of our life: environment, traffic, medicine, office,
manufacturing, limited resources, energy, telecommuting, etc. This will become reality
by the shift of paradigms in the fields of technology (where the #1 challenge will be
low power electronics), economics (where the #1 challenge will be low-cost fabrication)
and education (where the #1 challenge will be interdisciplinary creativity). It is no
longer the device, it is the system that counts. And it will be in silicon.
Microelectronics and software will change the world more than any other
technology in history (12)
MASS PRODUCTION OF NANOMETRE DEVICES
ALEC N BROERS
Churchill College
Cambridge CB3 ODS
It has become clear that the benefits of miniaturising electronic devices will
continue until dimensions enter the nanometre region, that is, below 0.1 /im. This is
beyond the resolution limit of the ultra-violet (UV) projection method used to
manufacture today's integrated circuits so a replacement will have to be found. The
major challenge for the new method is not resolution but the need to meet the extreme
image complexity and accuracy requirements at acceptable cost. Microchip images
already contain more than a billion pixels and if nanometre devices are to be cost-
competitive, the pixel count in a single 'chip' will have to approach a trillion. A pixel
is defined here as being four times smaller than the minimum feature size. Acceptable
cost is essential if there is to be commercial justification for further development. The
requirements of cost and pattern complexity will remain no matter what device
technology is used. The only way to avoid them would be to have the structures self-
assemble but as yet there are no methods for making contact to self-assembled structures
nor are materials that self-assemble suitable for electronics devices.
It is not possible to predict precisely the resolution limit for UV projection but the
ultimate camera might operate at a wavelength (A) of 160 nm and have a lens with a
Numerical Aperture (NA) of 0.75. Using the standard expression for the minimum
23
5. Luryi et al. (eds.), Future Trends in Microelectronics, 23-34.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
24
feature size, kA/NA, where k is typically 0.8, this camera would reproduce 0.17 /urn
lines, k depends on the mask type, the illumination and the resist contrast. With phase-
shift masks and off-axis illumination as described for example by Okazaki1, combined
with proximity-effect correction and a resist in which only a shallow surface layer needs
to be exposed2,3, a value of 0.5 might be reached for k and the feature size reduced to
0.1 /urn. This would allow the IG bit chip and its equivalents to be made with UV
projection and perhaps the 4 Gigabit generation as well. The depth of focus, A/(NA)2,
would fall to 0.28 /mi, but methods may be available for effectively increasing this, for
example, by using dual masks spaced along the optical axis.
In addition to small depth of focus, difficulties will remain with sources and with
the fabrication of optical components for the sub-200 nm region. Of these, densification
and colour centre formation in fused quartz, and the lack of suitably reliable laser
sources, have already been identified. The performance of lasers in the vacuum UV
regime will also have to be improved.
The most obvious way to penetrate the 0.1 /urn barrier is to retain the basic
concepts of UV projection but to use much shorter wavelength radiation. Optical
components for wavelengths below 100 nm are difficult to make but reasonable
performance has been obtained with multi-layer mirrors at 13 nm and this has led to
proposals for projection cameras4'5'6. The major difficulty is that the mirrors must be
fabricated with an accuracy of a few tenths of a nanometre and which is impossible
today but may be achievable in the future.
With X-ray lithography, the mask and wafer are separated by a gap that is large
enough to prevent them touching each other but not so wide that image degradation due
to diffraction is unacceptable. The minimum linewidth is approximately given by the
simple expression \/Ag, where A is the average wavelength of the X-rays and g is the
gap between mask and sample. -/Ag is the linewidth at which the intensity in a narrow
line first reaches that of the background with a transparent mask, assuming single
25
wavelength, coherent, illumination. In practice, the situation is more complex because
the source produces a spread of wavelengths, the absorber is not completely opaque and
the substrate is not completely transparent. In fact both introduce phase shifts and there
is the possibility of using these phase shifts to improve image contrast. Despite this
complexity, Vkg still gives a useful estimate of minimum linewidth. In the limit for
chip lithography, a gap of 5 /urn might be acceptable which, used in combination with
a wavelength of 1 nm, would permit 0.07 fan dimensions to be reproduced. In the
laboratory, intimate contact between mask and resist allows dimensions of a few tens
of nanometres to be reached but this would be impractical for production.
The only source that provides an adequate flux of X-rays is the electron
synchrotron storage-ring. Electrons in a storage-ring emit electromagnetic radiation as
they are constrained in their circular orbit by the magnets that comprise the ring. Tens
of wafers per hour can be exposed with the X-rays produced from a few degrees of the
360° orbital path of a few hundred milliampere, 500 MeV to 1 GeV, beam. The
divergence of the fan of X-rays leaving the ring is so small that the beam may be
considered to be parallel and the placement errors, that occur because the source is
divergent, become negligible.
Electron beams are used successfully for three applications; mask-making, 'quick-
turn-around' production of ASICs (Application Specific Integrated Circuits), and the
fabrication of devices in research and development. As already mentioned, they are too
slow and expensive for mass-production. Electron beams provide higher resolution than
any other method except local probe atom manipulation, allowing the size of devices
to be reduced to the practicable limit of resist processes, which at present falls at about
10 am.
To increase the throughput of scanning electron beam systems, methods are being
explored in which more and more of the pattern is exposed at each beam flash. The
slowest systems write the pattern one pixel at a time with a round electron beam whose
diameter is equal to the pixel size. This is very flexible and accurate but is extremely
slow. The next fastest systems use a beam which can take on any shape up to about
four times the minimum feature size9'10. This allows tens of pixels to be exposed at
each flash and makes it possible to manufacture ASICs. Finally, pattern cells
containing tens of shapes and thousands of pixels are projected in a single flash in what
are called 'character' or 'block' projectors. These are the only scanning systems
capable of potentially reaching realistic throughput for pattern replication11'1 ' .
Figure 1 shows estimates of the throughput (as a function of beam current density)
for the three types of scanning electron beam systems scaled to 0.05 ftm dimensions.
The following parameters were used in the estimates; minimum feature size 0.05 ftm,
pixel size 0.013 fim, pixel delay time for round beam 0.5 ns (this will only be possible
if the pattern is treated as a series of shapes with > 20 pixels per shape so that the delay
per shape is > ~ 10 ns), shape and character/block delay time 30 ns for shaped beam
and character projection systems, sub-field delay time 3 fis (it is assumed that there are
only two levels of deflection although three levels will probably be needed to meet the
required accuracy), chip delay time 100 ms, wafer change time 20 s, fraction of chip
area exposed 30% for round and shaped beam cases and 100% for character/block
exposure. Because the assumptions about electronic delay times are optimistic, the
throughputs in all cases are dominated by writing time and not by overheads and must
be considered to be higher than are likely to be achieved in practice.
The most serious problem with character/block projection is that the pattern must
be printed with a finite number of characters or blocks and yet customisation of design
rules to accommodate peculiarities of lithography systems has been considered
unacceptable in the past. A further difficulty is that proximity effects, which are
discussed later, can only be corrected by changing the shapes of the individual pattern
elements. The exposure dose has to be the same for all elements in a character or
block. Proximity effects are more easily corrected with single beams where it is
possible to change the dose as well as the shape.
10 . 1 1 111 1 10
D
5 0.1 _ Variable Shaped Beam ..-•"' -. 0.1
25 pixe s per s hape .-•''
o
CM
0.01 r 0.01
CL
Figure 1. Throughput for 0.05 fim scanning electron beam lithography systems.
An alternative to exposing many pixels at each beam flash is to use multiple beams.
Two distinctly different approaches are being taken to implement this idea. One is a
version of the block projection concept in which the character mask is replaced by a
multi-aperture mask that in effect produces an array of about 1000 beams14. The other
is to use ten or more miniature electron beam columns to write on a single wafer. Each
28
column has a field emission or thermal field emission cathode and addresses several
chips15. There are practical difficulties to be overcome with both approaches. With
the large array of beams there is the difficulty of building a multi-deflection unit that
can blank and deflect each beam individually and at the same time keep them precisely
registered with respect to each other. The electronic drive unit for this multi-deflection
unit must produce relatively high voltages and operate at very high speed. Control of
the beams will be difficult to achieve as the degree of electron-electron interaction in
the column will change with the number of beams that are active. The miniature
electron optical columns promise very high electron optical performance because of
their extremely short focal length, very low aberration, lenses but severe mechanical
tolerances will have to be met to achieve these low aberrations. It will also be difficult
to keep the field emission cathodes emitting uniformly and to fabricate the dense array
of electron detectors that will be needed to keep track of the individual beams. Overall,
there are many problems to be solved before either approach will be a contender for
production lithography.
The final possibility for achieving competitive throughput with electron beam
exposure is to build the electron equivalent of the UV step and repeat camera. This was
first attempted in the early 1970s16. The resolution of the projected image for small
field sizes can approach that of an electron microscope, that is a few tenths of a
nanometre, and the demagnification can be high enough to make mask fabrication
practicable. The first systems demagnified the mask 10 to 20 times. Thin metal stencil
masks were used in the early systems and it was proposed to resolve the difficulties with
unsupported and/or fragile features by using two masks for each level. More recently
a very thin supporting membrane has been used and the electrons that are scattered by
the membrane are removed by the projection lens stopping aperture17. Provided the
membrane is thin, that is less than 0.1 /mi, and the accelerating voltage is kept
relatively high, for example > 100 kV, the loss of current due to scattering can be
reduced to about 50%.
The use of a membrane solves some of the mask difficulties but the scattering
reduces the target current density and means that the beam current in the upper column
has to be increased thereby increasing electron-electron interactions. As with the shaped
beam scanning systems, the increase in energy spread blurs the image and this effect
will ultimately limits the exposure rate. Preliminary estimates suggest that the exposure
speed should be adequate, at least for dimensions down to 0.1 /an, but below 0.1 pm
this difficulty may be insoluble.
The absorber for the mask can be a relatively thin (<0.1 /mi) layer of a high
atomic weight material such as gold that is easy to fabricate. This is a significant
advantage over X-ray lithography where the absorber must be up to 0.4 fim thick.
29
A key difficulty with electron beam projection is pattern distortion. Distortion
arises in the mask and in the projection imaging. It should be possible to keep mask
distortion at an acceptable level through the use of supporting ribs which can be
fabricated with silicon anisotropic etching, but image distortion will have to be
corrected. It is proposed to accomplish this with dynamic corrections. The mask will
be illuminated with a small beam that will only cover a fraction of the image.
Distortion will then be corrected by tilting this beam as it is scanned over the mask to
complete the exposure. Variable axis condenser and projector lenses will have to be
used and the signals to the correction coils in these lenses will have to be extremely
accurately synchronised. The size and current in the illuminating beam will be as large
as electron-electron interactions allow.
7. Ion Beams
Lithography systems that use ions have been investigated for many years but there
are a number of factors that place them at a disadvantage compared with their electron
equivalents. Firstly, they have to use electrostatic lenses which have higher aberrations
than the magnetic lenses used with electrons. It is also not possible with electrostatic
lenses to overlap focusing and deflection as can be done with magnetic lenses and coils.
This overlapping has been found to be essential with electron beam systems if adequate
field size and beam aperture are to be obtained. The disadvantage of the higher
aberrations is offset somewhat when diffraction limits are encountered because of the
shorter wavelengths of the ions but this is unlikely to allow te performance of Over-
30
There is one significant advantage and that is that exposure with ions does not
suffer from the deleterious proximity effect encountered with electron exposure. Ions
do not penetrate the substrate deeply and hence are not backscattered through the resist
into areas away from their intended point of exposure. However, the lack of
penetration means that the resist is not exposed uniformly throughout its depth unless
high (> 100 kV) accelerating voltages are used. If the energy of the ions is increased
until the ions do penetrate through the resist into the substrate then unacceptable damage
of the sample may occur. This is not a problem for mask writing, and for direct-
writing with surface imaging resists, but to date few have considered the advantage of
proximity free exposure to be significant enough to override the shortcomings of ion
optical systems.
The ultimate resolution of ion optics is more than 20 times worse than it is for
electron optics, ~5 nm versus ~0.2 nm because of the higher aberrations of the
electrostatic lenses and the larger energy spread of the sources. This difference is only
marginally important for conventional resist exposure, as the resolution of resist itself
is about 10 nm, but for inorganic resists, where dimensions well below 10 nm can be
obtained, the ion optical resolution limit will become the fabrication limit.
Ion beam systems have been shown to be valuable for mask and circuit repair,
particularly for the removal of unwanted material, but for direct-write lithography the
systems proposed do not seem to offer any advantage over their electron equivalents
either in scanning or projection configurations.
8. Local Probes
The likelihood that local probes will prove useful for mass-production of nanometre
scale devices would therefore seem remote. For scientific work, however, their unique
ability to manipulate single atoms makes them ideal for fabricating simple test
structures. For example, Eigler and his coworkers have created 'quantum corrals' by
placed single atoms in a ring with an accuracy of less than a tenth of a nanometre19.
This remarkable resolution could not have been approached by any other method.
9. Conclusions
Electron projection methods can meet the resolution requirements easily but the
system complexity required to achieve practicable throughput will probably make this
approach too expensive and in any cases such systems will take years to develop. Ion
beam methods will be even more expensive and difficult with the difficulties
out-weighing their only advantage, which is lack of long-range proximity effect. In the
very long range, X-ray projection cameras may produce 0.05 //m capability but it may
take decades to perfect the methods needed to fabricate the mirror lenses. Multi-
electron beam columns may also offer a solution but precise control of hundreds of
beams presents a multitude of challenges.
Table 1 summarises the resolution limits of the different lithography methods. The
'Practical Limit' is the limit that should be achieved in production, and the 'Ultimate
Limit' is that which may be achieved in the laboratory under ideal conditions. Table 2
compares the different electron beam nanofabrication and nanostructuring methods.
The work discussed in this paper has been described by many workers in
publications that are too numerous to cite comprehensively. Further reading is available
32
in the proceedings of the International Symposia on Electron Ion and Photon Beams20
published each year in the Journal of Vacuum Science and Technology, the proceedings
of the International Conferences on MicroProcess published each year in the Japanese
Journal of Applied Physics21 and the proceedings of the Microcircuit conferences
published by Elsevier22.
10. References
-f
a, >. „
x>•" o
^ « « 8
v — —
>> A -° £ o 3- J3
>> <= 60
e .op 1 M
U Ui T-H
= 13 op =3 8 * X X
3 §
S c
«
- O 11
Sop 3 u 2 u p
s «§■ I E S u n ~ a e ii ts ■O
S
O -a
Z 2 al 3 ^
Z I ill
>> <u
CO
2* S 3
3 O. e 3
If 3. a.
g M
E a
[/I
3 »„
5 3 to
«=> 2S •ft ,<U
■ ^2 F, * 1 E3 r« &o
s e S * -0Ü « .3>
■i- * .s X S X
o §• ■* s.
o iH o 00 & ptio- '5b «-
O O" o o
n< o -o § s ° V
3 =* = "3
O. a H
D. E
rj a D. °
JJ oo 2 2?'3 ! •& a
3£ * ! ^ «
*5s go a-
«a ? i •a 5 ' II
>/"> _ El*?
•aB so «€ I
u o E
M .2-
, 'S 3
c Ja s «-
Is ^ oo
O vl
' d (Nj 00
3.5 &
i ä g"
1:2 „
i oo a
d "ob d öo E . d <s %
O n
<-i .SP 2
d M S ■ —
' uoo -
>^
° CM
o Z • o
iH O
1 ( <D O
2 o
2
O rH Z -S
1 J3
d
60 g
X
Ä -°
>
o
OH
Table 1. Present and ultimate capability for systems under development for chip
lithography.
34
LITHOGRAPHY
PRACTICAL LIMIT ULTIMATE LIMIT
TYPE
SERGE LURYI
State University of New York at Stony Brook
Stony Brook, NY, 11794-2350 USA
1. Introduction
In this work I shall discuss a new device fabrication principle which I would like to
refer to as active packaging (AP). The meaning of this term is that certain essential
fabrication steps (lithography, etching, metallization, etc.) are performed after the
partially processed device or circuit is packaged onto a host platform.
One of the most important goals of the AP concept is the combination of dissimilar
materials (notably, III-V compound semiconductors) with silicon integrated circuitry
(IC) on a single Si substrate [1]. This goal, now widely recognized as an important
research direction in microelectronics, is shared by other emerging technologies, such
as those based on heteroepitaxial and thin-film transfer techniques [2]. At the same
time, AP widens significantly the class of device structures that can be manufactured.
Our ultimate goal is not only to "teach the old dog new tricks" but also to greatly
expand the assortment of tricks available.
It is worth stressing that the word "packaging" is used somewhat unconventionally in
the AP context. Active packaging is a device fabrication technique, intended to
implement devices on a foreign (not necessarily even semiconductor) platform that
perform better than conventionally fabricated devices on their natural semiconductor
substrates. In many instances, AP enables the implementation of structures that
cannot be realistically obtained in another way, such as those requiring lithography on
opposite sides of a thin semiconductor film.
The principle of active packaging will be illustrated in the instance of a
heterostructure bipolar transistor (HBT) structure, schematically shown in Fig. 1.
Such a structure would reduce the parasitic capacitance between the base and the
collector electrodes, enabling ultrafast operation with oscillation frequencies in the
range of 300-400 GHz and even higher. This in turn would open up the possibility
of implementing on-chip millimeter-wave phased-array antenna systems [3].
35
S. Luryi et al. (eds.), Future Trends in Microelectronics, 35-43.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
36
collector
5 n+ InGaAs collector
4 n~ InGaAs subcollector
+
3 p InGaAs base
n InP wide-gap
2 emitter
1 n+ InGaAs emitter
contact
0 InP
substrate
Top side processing includes etching of the collector stripe down to the base layer,
evaporation of self-aligned contacts to the base, deposition of a passivating dielectric,
etching of via holes and metallization. At this point all the base and collector
contacts are connected in a circuit with lines running over the passivating dielectric.
The circuit is then covered by another ("interlevel") dielectric layer and a relatively
small number of selected points of the circuit are connected to "top" metal pads
through a second set of via holes. The top ("communication") pads may be relatively
wide (e.g., > 100 |xm). The interlevel dielectric (e.g., polyimyde) may be planarized.
No attempt is made at this stage to contact the emitter.
Flip-chip mount; "consulator" film. The circuit is then mounted on a "carrier" wafer
which has a mirror pattern of metallic communication pads. The carrier may be any
substrate, including glass, ceramics, etc., but first and foremost a silicon wafer that
has already undergone the integrated circuit processing. Connection between
communication pads is established with the help of an anisotropically conducting film
with electrical properties, illustrated in Fig. 3. The film must provide a short between
overlapping contacts and an open circuit otherwise. Such vertically conducting and
laterally insulating films, which may be called "consulators", can be prepared in a
variety of ways. The primary purpose of the consulator, besides providing vertical
electrical connections, is to provide a stable mechanical support for the packaged chip
- support that will become crucial when the InP substrate is removed.
t -10(im
TEES
base
MMlMlVlMM^iM
collector
3TT
TL
LÄ:::::£M*Mi£üi£
I I
to Si circuitry
Substrate removal and back-side processing. It is quite possible to etch the entire
InP substrate down, stopping at a 0.1 urn InGaAs layer. This step is based on the
well-known extreme selectivity between the etch rates for InP and InGaAs in
hydrochloric acid solutions. It is essential that the uncovered surface of layer 1 is
uniformly flat, adequate for performing a fine line optical lithography. A large hole
etched from the substrate side would not do, because there would be problems with
focal depth. To make lithographic alignment to the base contact level, the contact
metal should be seen with a sufficient contrast through layers 1-3. If this proves to
be inconvenient, then special topography features for the back-side alignment must be
provided at the top-side processing stage.
The final assembly is illustrated in Fig. 5. The emitter contact is established by a
standard lift-off evaporation of a suitable metal. It is well known that ohmic contacts
to n+ InGaAs are good without alloying. No elevated temperature procedures should
be contemplated after the chip has been mounted, because of the limited thermal
stability that can be expected of a consulator film and the need to preserve the
integrity of fully processed Si integrated circuits on the carrier wafer.
3. Advantages
The reduced base-collector capacitance offers significant advantages for microwave
performance of HBT. An enhancement of the maximum oscillation frequency f^
by a factor of 2 to 3 has been predicted [5,6] over optimized collector-down
structures. Moreover, with a suppression of the extrinsic collector capacitance Cö it
becomes possible to implement HBT structures with coherent effects in the base
[7,8], resulting in a power gain above the conventional cutoff frequencies. Figure 6
shows the modeled microwave characteristics of a collector-up HBT, in which the
base bandgap is graded so that the total base propagation delay x is much shorter than
the diffusive delay in a flat base of the same width [8]. The magnitude of the base
transport factor a= |a| exp(-27ii/x) decreases so slowly with increasing
frequency / that it becomes feasible to activate transit-time resonances far above
/T = 1/27IX. The fundamental peak in U occurs near 7t/T [7,8].
The coherent transistor can be designed to have the first high-gain peak at any desired
frequency, provided the effect is not destroyed by the parasitics. For frequencies
below 100 GHz it is possible to use conventional structures (Fig. 6a), where typically
the extrinsic (parasitic) collector capacitance C^ is about twice the intrinsic (useful)
capacitance Cc. In order to push the peak into a sub-millimeter wave range (Fig. 6b)
it is essential to reduce C^ below Cc. The AP process opens a way to substantially
reduce the extrinsic capacitance.
40
10 100
Frequency, GHz
Fig. 6a. Common-emitter current gain \h2i\ and the unilateral power gain
\U\ of a model coherent transistor with a special graded-gap base design,
optimized for stable oscillation at 94GHz. Base total width W=l\im.
Transistor is assumed loaded with the parasitics with state-of-the art equivalent
circuit parameters, e.g. CCX = 2CC. Conventional current-gain cutoff is
/T ~ 32 GHz, however the transistor also exhibits a range of current gain
| h2i | > 1 at / = 2nfT (near the second peak in U). The fundamental peak in
U occurs near It/T (after Refs. [7] and [8]).
40 \l I
* \
\ \ 300 K
\\ \\
30 — \\ \\ _
\\ \\
\ \ \\
CD
■D
\* \\ ....
c 20 \ |U| -
\ M
o \ \ 1
10 \ |h2,l2 \ _
HBTby \ VI
Active Packaging » I
c« < cc \ ll
\ "
0 —^—J —JJ— 1
10 100 1000
Frequency, GHz
metal
4. Applications
Based on AP HBT technology it is entirely feasible to implement local oscillators and
amplifiers that operate at millimeter and even submillimeter wavelengths. One
obvious application of such devices would be for satellite communication systems in
the atmospheric transmission window of 345 GHz.
Another extremely attractive application [3] is the possibility of fabricating
millimeter-wave phased arrays on a silicon chip. A X/2 spaced linear array of 20
elements radiating at 300 GHz would be about a centimeter long. The advantage of
having transistor oscillators is that the millimeter-wave beam can be electrically
steered off broadside by controlling the relative amplitudes of different oscillators,
while their relative phases are locked together by the evanescent wave interaction.
The point is that most available phase shifters used in centimeter wave phased array
systems are bulky elements that cannot be used in on-chip designs. Instead, we
should use electronic beam steering by controlling the amplitude of constant-phase
42
array elements [10]. As far as I am aware, this idea has not been employed in
practical phased-array antenna systems, perhaps because at centimeter wavelengths it
is more efficient to control the relative phases of array elements. In the millimeter and
submillimeter wavelength range amplitude steering appears to be the only realistic
way to implement purely electronic beam steering. Three-terminal devices are ideally
suited for this purpose. On-chip focal plane antenna arrays should have important
applications as steerable radar systems in avionics, automated manufacturing, and
especially in automobile collision avoidance and early warning systems.
5. Conclusion
Generality of the active packaging principle transcends microwave transistor
applications. The new degree of freedom in manufacturing - lithography on opposite
sides of a thin film - permits the implementation of a variety of new devices and
functions. Of the many possible examples, let me mention the possibility [11] of
fabricating a collector-up charge injection transistor with the channel-defining trench
etched in side 2 and aligned to the collector stripe on side 1. To indicate the scope
of contemplated applications, let us note that the technique makes feasible an active
directional coupler in which two edge-emitting laser resonators overlap in a portion of
their length. It also simplifies many schemes for integrating electronic and photonic
devices into a single functional unit to be placed within the integrated circuitry on a
silicon chip.
Vertical cavity surface emitting lasers (VCSEL) should have an important role in this
program, because such elements enable interchip communication directly from the
chip interior. Active packaging technology offers several advantages in the
integration of VCSELs with silicon VLSI. For example, it allows to use non-
epitaxial Bragg mirrors (such as stacks of ZnS and Si02 layers) not only for the top
but also for the bottom mirror of a VCSEL cavity. Also it permits the
implementation of tandem systems in which one VCSEL (master) works under
electrical injection of carriers while the other (slave) is optically pumped by the
former.
I believe that most significant applications of compound semiconductor electronics
will be associated with its use in silicon electronics. In terms of the old debate on Si
vs GaAs, my view is that silicon is the ultimate customer for GaAs. The logic of
industrial evolution will motivate new paths for a qualitative improvement of system
components, other than the traditional path of a steady reduction in fine-line feature
size. The principle of active packaging, illustrated in the present work using the
instance of implementing ultra-high performance InP HBT on a silicon chip, will
become one of the central design principles of future microelectronics.
43
References
Abstract
1. Introduction
The economics of the semiconductor industry are such that increases in productivity
have come at a rate experienced by no other industry in history except perhaps those
transformed by the advent of the steam engine. This exponential rate of productivity
growth has been driven by the availability of new generations of lithography tools
which have ratcheted down device ground rules (G/R's) at regular intervals.
An important ingredient of this scenario is the capacity of VLSI technology to
show overall performance improvements with the trend toward increasing
miniaturization. A key component which has not benefited from this miniaturization is
VLSI interconnect technology, which is becoming increasingly hard-pressed to deliver
device performance in the face of its own inability to provide performance improvement
through scaling and increasing complexity brought on by ever increasing circuit
densities and chip sizes.
These problems are not new, but they are becoming debilitating to VLSI designers
as microprocessors approach 3 ns cycle times [1]. In fact, much of the proposed
solution has been perceived for many years [2]. The sobering fact remains that nothing
short of a revolution in VLSI technology will provide a solution to the interconnection
dilemma. New materials such as copper wiring and low-epsilon dielectrics can and do
provide some relief, but a renewable source of performance improvement for VLSI
interconnects is not on the horizon. Our intention is to examine current interconnect
45
S. Luryi et al. (eds.), Future Trends in Microelectronics, 45-56.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
46
technology in light of trends occurring and expected to occur within the semiconductor
industry and to discuss the potential of those alternative and future technologies which
may relieve some of the current wiring burden.
Much of the robustness of current VLSI technology can be credited to a very solid and
extendable multi-level metal (MLM) technology. While this technology has benefited
immensely from rather recent breakthroughs in planarization (e.g. chemical-mechanical
polishing), many of its components are old stalwarts that have been reengineered
through several generations.
Some of the attributes that have benefited the Al/Si02 interconnect technology are
generally low line resistance, highly stable and rather low permittivity dielectric, metal
electromigration resistance, and overall compatibility with device processing
limitations.
called a TACT-TiN process in reference to the materials set involved: Al(Cu) alloy line
metallurgy (AC), Ti layers (T) over and under this line for electromigration resistance,
and TiN antireflective coating. The process summarized in Table 1 also includes
tungsten studs lined with PVD Ti/TiN, and a PECVD TEOS SiC>2 dielectric.
It should be pointed out that the rather long sequence of process steps associated
with the line/stud portion of the build may need to be repeated 4-5 times for a high
performance microprocessor design. It is quite easy to understand where the cost and
yield loss due to the BEOL originates.
Much of what is present in today's VLSI wiring technology is perfectly adequate for the
bulk of designs. As microprocessors become more and more pervasive in consumer
goods this trend will increase. It is only with the highest performing microprocessors,
those limited to a very small percentage of the overall application space, that we run
into performance difficulties due to the interconnect technology. Unfortunately, or
perhaps not, most semiconductor manufacturers attempt to cover all of their design
space with a single technology, since semiconductor technology is fairly expensive
49
either to license or develop. For companies with a very narrow range of products, this
may not be a problem. For companies with large product scopes, this situation
typically results in a high performance technology being developed which is eventually
amortized by incorporation into lower performance chip designs. The following
discussion will treat such a high performance technology and will emphasize those
technology elements which act as bottlenecks to overall system performance.
Even in the best of circumstances, we have seen that ideal scaling results in RC delays
that are proportional to S2Sc2. This loss of performance due to scaling occurs purely
from line length increase and cross-section reduction. The really bad news is that today's
interconnect technologies, which contain parallel path conductors (cladding) for
redundancy to safeguard against metal voiding, do not even scale this well! This is
simply the result of the interconnect dimensions decreasing to the point where the
resistivity of the cladding layer itself becomes important. Figure 2 illustrates the
cladding situations encountered for both aluminum and copper wiring and the effective
resistivity that results as G/R's are shrunk. The curves labeled "future" are predictions of
where the technology might end up as new or thinner cladding layers are implemented.
25
(1) 20
F
F
o
ü 15
>>
ü
o
a>
10
c
<u
u
a. 5
II
1985 1988 1991 1994
Logic Technology Generation
As Fig. 3 would suggest, the fraction of the overall microprocessor cycle time that is
attributed to wire delay has been increasing at an alarming rate over the past decade. The
cause for this rapid rate is obvious. As already discussed, interconnect RC delays are
increasing as S2Sc2 with each successive technology. At the same time, device
51
While not very innovative, it is the belief of the authors that there is no white knight
on the horizon to alleviate the VLSI wiring bottleneck. Several novel future
technologies which provide hope for the very distant future will be mentioned in the
following section. But to address the requirements of the next couple of generations of
CMOS microprocessor design, we need to provide several shots of technological
adrenaline, and soon!
2.0
' !""
■ ;
1.6 "i
■ i
1.0
■ I
■ !
■ I
0.5 -!
■ I.
- j.
■ V
J
J
Al/Ox Cu/Ox AI/PI Cu/PI Cu/Ox Cu/FP Cu/PI Cu/FP
scaled scaled scaled
Technology
While materials changes can be used to provide performance improvement, they cannot
provide sufficient RC delay improvement to treat the long lossy line problem. An
obvious solution presents itself, that of providing additional levels of unsealed wiring to
specifically treat those circuits which are potentially cycle time limiting [1]. These
same "fat wire" levels can also by used for power bussing and clock distribution as is
currently done with the upper levels. The distinction between "fat wire" levels and what
is currently available is that "fat wires" have more cross-sectional area. In practice, this
is a fairly difficult structure to fabricate. Metal RIE is particularly unforgiving when it
comes to etching metal several microns thick.
Fortunately, copper damascene processing may provide an easier route to
fabrication of these large conductor structures [5]. Indeed, many of the types of
structures required have already been demonstrated using copper damascene technology.
Further discussion of the utility of unsealed wiring levels can be found later in this
volume [6].
53
Another solution to the long lossy line problem, but one which is much more
controversial and slightly less elegant, is to provide long cross-chip wiring in the form
of interposer wiring on a single- or multi-chip module. Such solutions will be
particularly attractive for multi-processor designs.
Current packaging technology is available to effect such a solution so the real
decision is as much an economic one as it is a technological one. Large chips (greater
than 15 mm) can benefit by as much as 20% in RC delay by going to thin film wiring
as opposed to using double-wide 0.35 G/R on-chip wiring. If the die size plateau is not
reached, as we are projecting, the use of alternative low resistance, cross-chip wiring,
such as is available in the form of thin-film wiring on SCM's or MCM's, will be a
viable alternative to additional BEOL wiring levels.
6. Future Technology
Very few viable technologies which directly impact the VLSI wiring bottleneck loom
shimmering on the horizon. There are glimmers of hope, such as room temperature
superconductors, which might make things interesting, but if we ground ourselves in
reality and make critical evaluations of what is available, the undeniable conclusion is
that materials technology is the only relatively near-term remedy for the interconnection
bottleneck, and that remedy only treats the symptoms. Some of the more exotic
technologies which offer some hope are the following.
Several materials alternatives for VLSI wiring processes have already been discussed.
The purpose of this section is to address the feasibility of the more exotic forms of
physical interconnection available. Throughout the discussion we continue to consider
54
the current Si-based VLSI technology as the base with which any conceived alternative
must compete and ultimately displace.
6.2.1 Optical.
At this time there is no implementable scheme for optical interconnects in Si-based
VLSI technology. The more unfortunate point is that there is currently no obvious path
to implementation or that if a path was available, whether it would be practical [8]. One
might state that on-chip (Si) interconnects are not in the realm of reality for optical
interconnects and be pretty safe. It is apparent that the more easily obtainable and,
perhaps, more useful role for optical interconnects is as intra-MCM and, in the future,
as inter-MCM connections.
6.2.2 Superconducting.
For many years after the discovery of the high temperature superconducting cuprates, the
goal of producing usable on-chip (Si) interconnects was viewed as highly desirable.
Such an implementation still has many hurdles to cross [9], since many of the high
temperature superconductors are metastable phases in their own right, and more
seriously, since they react quite quickly with Si, SiC>2, polymers, etc. at the
temperatures required to deposit them in thin-film form. Another serious limitation is
their inability to carry the high current-densities required of on-chip interconnections.
Unless new discoveries occur, the usage of high Tc superconductors for on-chip wiring
appears completely out of the question.
One of the most exciting prospects for alternative computing which has major impact
on the wiring aspects of system design is the concept of quantum cellular automata
[10]. This concept is clearly many, many years away from usable products, but it does
contain some clearly valuable attributes.
Of these attributes, the most astonishing is that Coulombically coupled quantum
cellular automata do not require wires at all! Wires are constructed of the devices
themselves. In the most concrete suggestions for implementation [11], devices exist as
islands of nano-phase material patterned using an STM tip. While this fabrication
concept and the actual usage of such a cellular automaton scheme is clearly immature
[12], the ability to conceive of an implementation scheme is undoubtedly a step in the
right direction.
7. Summary
Modern VLSI technology is faced with a challenge in terms of its ability to reasonably
wire highly complex microprocessors and, at the same time, provide sufficient
performance in those interconnections to keep pace with steadily improving device
characteristics and system cycle times.
The current generation of VLSI wiring technology is superb. It has been
reengineered through many generations and is still capable of delivering on both its
reliability and density requirements. Wiring delay as a percentage of system cycle time,
however, has been increasing. Today's highest performing microprocessors are clearly at
55
the mercy of wire delay constraints. Increasing chip sizes and decreasing G/R's are
tightening these constraints.
Unfortunately, the avenues available to alleviate the wiring problem are not
numerous. Materials substitution in the form of copper metallurgy and low epsilon
dielectrics can provide some relief; there is increased activity and emphasis in the
industry to adopt these. In addition, it is expected that chip sizes and overall machine
complexity (circuit count) will begin to plateau. Design methodology will begin to
concentrate on cycle time reduction which will drive chip size and interconnect length
down, but will also demand extremely low RC wiring paths (fat wires) in order to
reduce time-of-flight delays.
While several novel technological concepts exist to address the wiring issue, none
of these seem sufficient to impact the industry in the near- to medium-term. This being
true, the semiconductor industry is left with the options of introducing new materials
technologies which give one-time boosts to wiring performance, introducing unsealed
wiring to treat cycle limiting paths, and awaiting the arrival of less complex and smaller
chip-size microprocessor designs.
8. Acknowledgments
The authors would like to thank the following people for both technical assistance and
critical comments: Rolf Landauer, John Heidenreich, J. Frank White, John Hummel,
Steve Greco, C.-K. Hu, George Sai-Halasz, and Kerry Bernstein.
9. References
H. VAN HOUTEN
Philips Research Laboratories
Professor Holstlaan 4
5656 AA Eindhoven, The Netherlands
1. INTRODUCTION
From the end of world war II to the demise of the cold war, the expen-
diture on basic research in solid state physics and materials science has
often been justified by reference to the invention of the transistor, and the
resulting revolution in electronics [1]. It is the stated aim of this conference
to study the future of microelectronics, beyond the time when the shrink-
age of CMOS feature sizes will have come to a stop. While major western
industries are losing confidence in the business generating potential of ba-
sic research in physics and materials science, government funding agencies
are sponsoring research programmes on esoteric subjects such as quantum
devices, single electron tunneling, molecular electronics, atom manipula-
tion, or self-assembly [2]. In doing this, they are guided by the idea that
breakthroughs in these fields will shift the physical limits of the miniatur-
ization trend to the nanometer scale, and thereby ensure decades of con-
tinued growth for the electronics industry. The fact that this would require
"Revolutionary chip architectures which remove the current interconnec-
tion limitation to functional density", as well as "Revolutionary devices
which make use of physical phenomena on a much smaller scale than tran-
sistors" does not seem to deter the scientific community [3]. Indeed, similar
revolutionary ambitions abound in the fields of computing and data storage
(the all-optical computer, the biochip, scanned probe data storage,...).
This paper presents the view point that those of us trying to defend
the strategic role of research in physics and materials science should not
advocate such trendy goals, which remind us of failures from the past, for
example the Josephson computer [4]. This type of desired breakthrough is
57
S. Luryi et al. (eds.), Future Trends in Microelectronics, 57-70.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
58
indicated in Fig. 1 by a small cloud, situated at the intersection of a main
technology trend and its (perceived) fundamental performance limit. While
taking an ambitious - but naive and rather unimaginative - aim far beyond
such limits, these endeavours often have intrinsic flaws, do not fulfill all the
requirements, or do not address the real limitations or bottlenecks of the
existing solutions [5].
Instead, we ought to identify more realistic goals for our strategic physics
and materials research. Goals that make sense from a technical and business
point of view can be found, even in a world dominated by dreams of multi-
media systems, software, and services. To help assess such goals, it is useful
to distinguish between "trend feeding" and "trend setting" innovation. 1
The following definitions will be used in this paper.
- Trend feeding innovation is characteristic of an established prod-
uct/market combination. Support is given to an existing trend of im-
provement versus time of key performance parameters (data rate, data
density, instructions per second, packing density, cost per bit, etc).
It can be based on evolutionary or more radical technology changes
(see arrows in Fig. 1), but all technical and economical compatibility
requirements characteristic of the main trend must be met simultane-
ously.
- Trend setting innovation addresses a new product/market combi-
nation. The performance level of existing trends is often not met, at
least initially. One or two outstanding features of a new technology are
exploited, to create the initial market, and to set the new trend.
The nature of these two types of innovation will be described in more detail
in the next two sections. We also present some examples, which clarify the
difference, and which may point the way to some more general themes for
innovation based on physics and materials science.
2. TREND FEEDING
2.1. SYSTEM TRENDS
System engineers tend to think in terms of hardware options available to-
day. Yet, system trends can provide directions for hardware research. The
Japanese company NEC has been guided over decades by their vision of
the merging of computers and communication (C&C)[7j. Today, several
companies are convinced that a further merging of these two product areas
with consumer electronics (CC&C) is imminent. Technical driving forces
behind this unification are digitalization, and the need for interconnect-
J
In an interesting paper, Bower and Christensen recently made a somewhat similar
distinction between disruptive and non-disruptive technologies[6]
59
time
Figure 1. Typical exponential trend in performance versus time, fed by evolutionary
or more radical technology changes (as indicated by arrows). According to conventional
wisdom, physics and materials science should create breakthroughs (indicated by the
small cloud) that would dramatically shift the perceived physical limits of the existing
technology (dashed line), thereby ensuring a continuation of the same trend for further
decades.
ing the traditional stand alone electronic "boxes" (such as a TV, a PC,
a telephone) using networks. Quite important is also the recognition that
a common layered system architecture may underpin quite different func-
tions. A service (a movie, or a telephone conversation) is distributed over
a general purpose utility (e.g. a cable network). The system is ruled by
an operating system, governed by (perhaps downloadable) software. The
hardware is typically build in a modular fashion, with a microprocessor, a
hierarchy of storage modules, a display function, a user interface, etc. A
fashionable concept on this emerging multimedia battle ground is that of
the value chain. In the trajectory from hardware component to service ren-
dered, each of the parties involved will strive for the most profitable slice
of the pie. While some see no compelling reason to remain active in hard-
ware, wishing instead to concentrate on the (more profitable) software and
services, others see interesting opportunities for "key components" or "key
60
modules", which constitute the core of the hardware (and which may have
added value through embedded software). The drive for increased function-
ality of key components is reflected in the slogans "a system on a chip",
and "a system on a display". The power balance between manufacturers of
components (e.g. a microprocessor) and set makers (e.g. a PC) is uncertain.
The vagueness of the demarcation line between components, modules, and
systems has a significant impact on packaging issues.
2.3. EXAMPLES
Radical technology changes, intended to feed main trends, have to meet
the same severe compatibility requirements (in terms of manufacturing,
standards, etc.) as more evolutionary ones - on top of a demonstrated
compelling advantage in one or several aspects. A boundary condition usu-
ally is that the increase in product cost should be negligible. Proving that
all of these demands can be met simultaneously is a major challenge for
industrial research, requiring a sustained commitment over an extended
61
Si
S1O2
Metal
Poly Si
PZT
Pt
period of time. Below, we will give some examples from the fields of em-
bedded memories, packaging, and magnetic data storage. These examples
were selected to illustrate four rather general themes for "trend feeding"
innovation based on physics and materials science:
— The enhancement of the performance or functionality of silicon based
microelectronics through new materials.
— Miniaturization enabled by sophistication in packaging.
— Opto-electronic technologies providing solutions for systems demand-
ing high performance in terms of processing (using electronics) as well
as communication (using photonics).
— The migration of LSI technology to application areas which are served
today by non-planar (i.e. bulk like) technologies.
HH o-WM^-o
Figure 3. Technology for an Integrated Components Module. From left to right one sees
an IC mounted upside down onto a Si substrate, containing a resistor, a capacitor, and
an inductor.
(L)
IC
'sense
Read gap
Figure 5. Schematic view of the play back part of two of the digital channels of a DCC
head. For clarity the left channel shows only the magnetoresistive element (with the
barber pole) and the lower half of the magnetic yoke. For the right channel the complete
read yoke is shown. In practice, a write yoke is processed on top of this.
will occur at even higher clock frequencies (exceeding a GHz in the near
future). This gives rise to significant electromagnetic compatibility (EMC)
problems, which may conflict with government directives. An interesting
approach is to integrate the optical waveguides, a laser source, modulators,
and detectors on the substrate of a single "multi-chip module", using flip-
chip bonding to connect the IC's [12]. Such a hybrid technology has the
advantage that standard IC's can be used. An alternative low-cost approach
under investigation in Philips Research is to add a few optical links to a
printed circuit board (PCB). In addition to the benefits mentioned before,
optical interconnects on a PCB could eliminate the need for multilayer
electrical interconnects. The PCB environment also provides enough space,
and conventional electro-optical technology (e.g. polymer waveguides) can
be used. This approach, which focuses on studying a systems concept rather
than a single device, is illustrated in Fig. 4.
3. TREND SETTING
3.1. NEW PRODUCTS, NEW MARKETS
hole
injecting layer-
electron
injecting layer-
plastic substrate
4. CONCLUSIONS
We have argued in this paper that strategic roles can be found for physics
and materials science, in support of continued growth of the microelectron-
ics industry. Instead of defending the mainstream vision that physics and
materials science should attack perceived fundamental limits of downscal-
ing by adventurously trying to invent yet another successor of the transistor
(which could be labelled as "trend fighting"), we have sketched two alter-
native routes.
Firstly, there is room for "trend feeding" innovation. This may be based
on fairly radical technology, but it must be compatible with existing tech-
nologies, and it must meet all requirements characteristic of the main trend.
Secondly, there is room for "trend setting" innovation. Here one should
not try to compete with the dominant trend. Instead, the focus should be
on a new product/market combination, based on one or two outstanding
properties (with a reasonable performance in other aspects).
We stress that these two types of innovation do not rely on technology
only: they also offer challenging subjects for physics and materials research.
Examples are the phenomenon of giant magnetoresistance in metallic and
oxidic multilayers, and the interplay of transport and luminescence in dis-
ordered polymer films. Last but not least, an entrepreneurial attitude is a
prerequisite for success in both cases.
5. Acknowledgement
References
1. Physical Review: Centenary-from basic research to high technology, Physics Today,
October 1993.
70
2. Engineering a Small World, From Atomic Manipulation to Microfabrication, Science
254 (1991) 1300-1342.
3. R.T. Bate, G.A. Frazier, W.R. Frensley, J.W. Lee, and M.A. Reed, Prospects for
quantum integrated circuits, SPIE Vol. 792 (1987) 26-35.
4. W. Anacker, Computing at 4 degrees Kelvin, IEEE Spectrum May 1979, 26-37.
5. R. Landauer, Advanced Technology and Truth in Advertising, Physica A 168 (1990)
75-87.
6. Joseph L. Bower and Clayton M. Christensen, Disruptive Technologies: Catching
the Wave, Harvard Business Review, January-February 1995, p. 43-53.
7. W. Aspray, Interview with Koji Kobayashi, in Engineers as Executives (IEEE Press,
New York, 1995).
8. P.K. Larsen, G.A.CM. Spierings, R. Cuppens, and G.J.M. Dormans, Ferroelectrics
and high permittivity dielectrics for memory applications, Microelectronic Engineer-
ing 22, (1993) 53-60.
9. O. Auciello and R. Waser, Science and Technology of Electroceramic Thin Films,
NATO ASI Series E, Vol. 284 (Kluwer, Dordrecht, 1995).
10. L.T. Canham, Appl. Phys. Lett. 57, 1046 (1990).
11. N.J. Pulsford, unpublished results.
12. T.E. van Eck, G.F. Lipscomb, A.J. Ticknor, J.F. Valley, and R. Lytel, Appl.Optics
31, 6823 (1992).
13. Y. Miura, Advances in magnetic disc storage technology, J. Magnetism and Magnetic
Materials 134 (1994) 209-216.
14. J.L. Simonds, Magnetoelectronics today and tomorrow, Physics Today April 1995,
26-32.
15. W. Folkerts, Magneto-resistive thin film heads for tape recording: past, present, and
future, Read/Write 18 (1994) 8-11. W. Folkerts, J.C.S. Kools, M.C. de Nooijer,
J.J.M. Ruigrok, L. Postma, K.-M.H. Lenssen, G.H.J. Somers and R. Coehoorn,
Performance of Yoke type GMR heads, Proc. Intermag. '95, San Antonio.
16. P. Ranganath Nayak and J.M. Ketteringham, Breakthroughs, Arthur D. Little Inc.
17. See, e.g., A.R. Brown, A. Pomp, D.M. de Leeuw, D.B.M. Klaassen, and E.E.
Havinga, Precursor route pentacene metal-insulator-semiconductor field effect tran-
sistors, Appl. Phys. Lett, to be published.
18. D. Braun, A.J. Heeger, Visible light emission from semiconducting polymer diodes,
Appl. Phys. Lett. 58, 1982 (1991); J.H. Burroughes, D.D.C. Bradley, A.R. Brown,
R.N. Marks, R.H. Friend, P.L.Burn, and A.B. Holmes, high emitting diodes based
on conjugated polymers, Nature 347, 539 (1990).
19. F. Gamier, R. Hajlaoui, A. Yassra, Pratima Srivastava, All-polymer field effect tran-
sistor realized by printing techniques, Science 265, 1684 (1994).
20. R. Friend, D. Bradley, and A. Holmes, Polymer LEDs, Physics World November
1992, p. 42; P. May, Polymer Electronics - fact or fantasy?, Physics World, March
1995, p. 52.
GROWING UP IN THE SHADOW OF A SILICON OLDER
BROTHER': TALES OF AN ABUSIVE CHILDHOOD FROM GaAs
AND OTHER NEW TECHNOLOGY SIBLINGS!
PAUL R. JAY
Microwave Modules Group
Northern Telecom Ltd.
P. O. Box 3511, Station 'C
Ottawa, Ontario, Canada K1Y4H7
1. Introduction
Not surprisingly, the gallium arsenide community has been, for some years, the object
of focused aggression and resentment from their colleagues in the silicon industry.
Family psychologists will tell us that such types of "sibling rivalry" are not
uncommon, and that ultimately the older offspring will come to accept and even value
the younger upstart that has created such a distraction within the semiconductor family!
Perhaps this change of attitude is beginning to happen as GaAs technology is
establishing itself as an authority in certain applications areas, and the silicon
protagonists realize that the new technology challenges only certain niche areas, as
opposed to undermining the whole basis of the silicon economy. Indeed, it is becoming
apparent that certain aspects of technology developed to cope with the particular issues
facing GaAs (such as rapid thermal annealing) can be usefully applied to Si processing.
Having established this platform for a provocative discussion in a workshop
atmosphere, this paper aims to summarize some of the recent applications achievements
of GaAs technology, and to examine the history of the emergence of GaAs to determine
what lessons can be learned from the pitfalls encountered en route. This allows us to
look for some sort of "selection rules" that might be useful to test a proposed new
technology and estimate its right to survival. The analysis also begs the question
presented by continued progress down the paths towards further miniaturization of
existing technologies for higher performance and greater speed: "Is a fundamentally new
approach to device technology required?"
In order to illustrate the fact that GaAs technology is now able to make useful
contributions to the electronics marketplace, this section will describe a few examples
of applications at different volume levels, many drawn from a workshop presented a few
years ago at the IEEE GaAs IC Symposium [1]. The range of successful applications
now runs from relatively low volumes of complex, high-performance IC's to substantial
71
S. Luryi et al. (eds.), Future Trends in Microelectronics, 71-86.
© 1996 Khmer Academic Publishers. Printed in the Netherlands.
72
volumes (in excess of millions of IC's per year) usually involving IC's optimized for
minimum size and cost. It is interesting to observe that whereas some of the earlier
applications fulfilled by GaAs involved very high speed digital functionality, the major
markets currently creating a renewed interest in GaAs are principally driven by the needs
of microwave functions in telecommunications and wireless transmission.
For many years now, Hewlett-Packard have been employing linear and microwave GaAs
IC's in their high speed test and measurement instruments [2]. The introduction of GaAs
chips was usually done on a case-by-case basis where functional benefit was evident
from the prototype demonstrations in terms of a simplified overall architecture and an
associated performance enhancement. The types of series volumes involved are not
necessarily large, but the value of developing an application-specific IC (ASIC) for one
product line can be justified even for product runs of 5,000-10,000 units, where an NRE
cost of $100k would involve only $10-20 additional cost per unit. In order to evaluate
and introduce these codes, HP sustains an in-house production line to supply their
commercial needs. Typical functions involved are microwave amplifiers, switches, and
high frequency dividers, and a number of the applications have enjoyed additional
advantages as simplifying retrofits in field service situations.
Building on the successful experience derived from programs such as the 'ALTAIR'
Wireless Ethernet project [3], Motorola have in recent years established a $100 million
GaAs IC facility handling 4" wafers, and are currently preparing to run 6" GaAs through
their process as substrates of the required quality become routinely available.
The ALTAIR system uses 18 GHz wireless links to provide up to 15 Mb/s of data
connectivity in a LAN (local area network). Key to the low cost and flexibility of this
application is the implementation of the microwave functions on GaAs MMIC
(monolithic microwave integrated circuit) chips operating at between 2 and 20 GHz. In
73
The whole field of GaAs technology in North America (and to a large extent also in
Europe) owes a major acknowledgment to the support received over the years from
military funding programs. Although recent years have seen a considerable swing away
from that type of application, there are still several companies pursuing the application
of GaAs to the perennial problem of cost-effective (and weight-effective) manufacture of
large arrays (around 3000 RF modules per face) for use in beam-steerable radar systems.
The advantage of this type of design is that each module embodies individually-
controllable transmit and receive functions, such that the focal point of the radar beam
may be steered electronically rather than mechanically, thus offering a much greater
degree of agility to deal with multiple hazards at different heights and distances.
Especially for airborne systems, the integration possibilities and performance
advantages of GaAs IC's enable the required functionality to be concentrated into a
minimum volume, and Westinghouse Electric have been able to apply their long
experience in all aspects of GaAs technology [4,5] to this subject. Among topics
currently being addressed in their planned evolution, the use of HBT (heterojunction
bipolar transistor) technology offers improved linearity and power efficiency across the
very large number of output stages in a given system. Recent announcements indicate
that Westinghouse is further enhancing investment in their GaAs capability to supply
commercial MESFET, HBT and p-HEMT IC's.
The supercomputer industry has had an interest in GaAs that focuses more on high level
integration of high speed digital functionality, and has met with rather more severe
challenges than the microwave side of the business. This type of need has driven work
aimed at high performance in a context of reproducibility and uniformity as required for
VLSI GaAs. Early in 1991, Convex Computer Corporation announced the development
of a 2 Gigaflop supercomputer based on GaAs IC's in combination with other high
speed technologies. By October of that year they had successfully delivered their first
C3800 machines based on an architecture using as many as 30 different GaAs IC
designs with up to 45,000 gates/chip. The Convex experience grew out of confidence
gained in GaAs substitutions for high speed Si bipolar IC's in earlier machines, which
enabled equivalent performance but for lower dissipated power. The IC's were realized
through a close and successful collaboration with Vitesse Semiconductor, whose self-
aligned gate (SAG) FET process with direct coupled FET logic dissipated the low power
levels that enabled the C3800 to be air-cooled.
In contrast to the air-cooled Convex machine, the CRAY-3 machine from Cray
Computer Corporation was based on an innovative packaging approach using total
liquid immersion of several thousand MSI (medium scale integration) GaAs chips for
74
effective heat removal. This architecture, together with the 100 ps gate delay of the
GaAs logic chips gave the CRAY-3 its 2 nanosecond clock speeds; more than twice the
overall performance per CPU as compared to its CRAY-2 predecessor. The first CRAY-
3 went into service in 1993, and work at Cray was focusing on more highly integrated
GaAs chips using a SAGFET technology for the 1 nanosecond clock CRAY-4, until
the recent announcement of the unfortunate demise of the Cray Computer Corporation.
At the time of writing, efforts are underway to sustain the 4" captive GaAs facility
developed at Cray, with a view to helping meet the current increased demand for
microwave GaAs.
Both the Cray and Convex programs have suffered from the drop in market share
of the supercomputer manufacturers rather more than from inadequacies of the GaAs
technology per se although it is clear that the demands of highly integrated GaAs digital
circuitry require a greater degree of process maturity and uniformity than is sufficient for
microwave IC's. In addition to this constraint, both of the supercomputer applications
pushed GaAs into a ground breaking domain of high-speed interconnect and packaging.
At least in the Cray case, the packaging issues rather than the GaAs capability seem to
have been the limiting factor in the commercial viability of the project.
The materials technology of the GaAs system is inherently more complex than that of
Si, and this inevitably also contributes to a process technology requiring control of
more degrees of freedom. GaAs technologists have learned many valuable lessons
concerning the material purity, its stability under thermal treatments, and especially the
sensitive nature of its surface and interface chemistry. Even armed with this new
knowledge, the Si technology still has a major advantage of many years of high volume
manufacture, and the authority and empirical understanding conferred by that history
carry considerable weight in comparisons with the III-V systems.
As an example, the Si circuit designer knows from detailed statistics that for a
given geometry, the performance of his design medium will follow a well-defined set of
rules, within quite close tolerances. Even the degradation of the tolerances at
progressively smaller dimensions is relatively predictable. The GaAs designer, on the
other hand, has to contend with a much smaller body of statistics for his prior data, and
even now is dealing with relatively small batches of "identical" substrates from each
boule. As a result, to ensure that his circuit will operate over the required range of
operating conditions and process parameters, the designer must expend extra performance
margin, power budget, and even real estate, to guarantee the objectives. This maturity
penalty for GaAs IC's is eroding more rapidly now as large volume applications become
established, but it has severely narrowed the predicted performance gap between GaAs
and Si for a number of situations.
Such an argument admits that GaAs is still on the learning curve, and this is both
good and bad news. It is good in that the performance margin against Si can be expected
to widen as volumes further increase, but bad in that a major area for improvement is
the relationship between starting materials and wafer processing technology, which is
all too often a "vendor-supplier" one, rather than a true team effort with open
understanding on both sides. Nevertheless, the performance/cost combination still leaves
much encouragement for GaAs, such as illustrated by the comparative extrapolations of
cost vs. yield for GaAs IC's by Skinner [6], who demonstrates the extent to which (for
large high speed IC's), packaging and testing dominate the cost much more than the
contribution from the cost of the processed wafers.
One of the key beneficial aspects of GaAs technology is the semi-insulating nature of
the high resistivity substrates which enables the close proximity of high frequency
components with low parasitics and minimum interaction between adjacent devices.
76
Assuring the stability of the insulating condition and especially its electrical
completeness at all frequencies, has been a major subject of study over the last 15-20
years. The unusual physics of this deep-level system has enabled the database to be
substantially enriched by research programs from many related disciplines and has
provided the materials growers with a wealth of models and solutions by which they
may understand and control the seemingly delicate semi-insulating situation.
Unfortunately, the subject of deep level studies became so interesting in its own
right that the burgeoning body of data eventually added little to the knowledge of how
best to control the material, and perhaps even started to cloud the issue. Ultimately the
solution evolved more out of practical adaptation based on statistical experiments at a
manufacturing level, than out of a detailed understanding of the processes involved.
A feature of this is that, even now, many different GaAs fabs use substantially
different "recipes" for dealing with implant annealing, backgating etc., and this
translates to different demands presented to the substrate suppliers, and thereby hinders
the convergence of requirements of the base material, and so sustains higher substrate
costs.
It is inevitable that during the parallel development of a new IC technology and the
circuit designs that will use it, some mismatch will occur between the needs of the
designers and the readiness or optimization of the wafer fab. At this point, especially in
a race for market share, the pressures to deliver can easily result in situations that are
less productive than "perfect teamwork", especially if the two groups are parts of
different organizations.
Many new product developers have commented from their experiences that the
notion of IC design on the basis of "Foundry PCM (process control monitor)
Parameters" is fundamentally flawed in that no PCM set can perfectly guarantee
operation of the circuit, and reciprocally, no foundry would wish to be constrained as
tightly as the spreads of an "ideal" PCM set would require.
The resultant compromise has to be an atmosphere of close cooperation between
designer and wafer fab, and this is most often achieved by the two functions being part
of the same organization. In many cases, development groups have found that a mutual
training activity in the form of in-house "User-design courses" serves both to sensitize
product and system designers to the new technology's capabilities and limitations, and
to create a shared sense of ownership in advance of the issues that are bound to arise.
Most of the original GaAs foundries have now moved towards a basis of more
"selected customers", and have recognized that the notion of "uncommitted ASIC multi-
project chip designs" tends to show a relatively low yield of true product developments,
whilst requiring substantial levels of engineering support. For this reason, most new
product developments for GaAs now take place either as customized ASIC programs,
(with the associated NRE costs) or as in-house developments in a vertically integrated
organization.
In spite of the fact that they are more proof against damage from radiation effects
(because of the absence of an oxide layer), GaAs devices and IC's are potentially
sensitive to ESD (electrostatic discharge) damage, as many manufacturers have
77
discovered by experience. An oft repeated adage is that "you don't realize that you had a
major ESD loss problem until you have installed all the measures to eliminate it!" This
is due in part to the latent nature of many forms of ESD damage, which will only show
at some later stage as a failure apparently due to something else. Operating practices and
disciplines need to be upgraded and constantly reinforced to eliminate this problem. A
typical example frequently encountered in system development labs is the use of heat
guns to either test the temperature sensitivity of a suspect GaAs part, or even to remove
it from the board. Since heat guns are an excellent source of charged ions, the
suspicions about the GaAs part were often proven apparently true as a result of this
barbaric treatment.
True reliability data on a brand-new technology can only be obtained after the
requisite number of hours of in-fhe-field operation on a statistically significant batch of
samples. Nevertheless, useful predictive data can be derived from studies based on
accelerated life-tests, although the variety of activation energies quoted for similar GaAs
technologies suggests that some subjectivity is still involved. In general, for GaAs
FET's using recessed-gate technologies (usually with Pt or Pd and Au gate electrodes),
the predominant failure mechanism over time is referred to as a "sinking-gate"
mechanism, and this is generally slow enough not to be an issue in most applications.
In technologies using refractory metal gates (such as SAG processes with WSi gates)
the ability of the gate metal to withstand the high implant temperatures means that it is
very stable at typical operating temperatures, and the failure mechanism of the Ohmic
contacts then appears as the next "layer of the onion".
More recent studies have highlighted an additional mechanism that occurs when
recessed-gate devices (using Pt or Pd gates) are exposed to hydrogen-containing
atmospheres in a confined enclosure. This can occur when a recessed-gate GaAs IC is
hermetically sealed in a Kovar package, since the outgassing of H2 from the walls of
the package is sufficient to provoke an effect [7]. Although this effect is now understood
and largely under control, its recent appearance highlights the need for exhaustive
reliability evaluation of new technologies before they are introduced to field
applications.
The spectacular operation of a new high speed technology is of little value if it cannot
be communicated to the rest of the system, and yet for the most part the packaging of a
GaAs IC is frequently treated as a design afterthought. Inappropriate termination of
digital input/output cells or poorly-matched RF ports on a microwave IC can seriously
degrade the simulated performance of the circuit, and both cases have caused GaAs
designers to address issues of high speed packaging necessary to deliver their
commodity.
Indeed, one of the strengths of a multi-chip-module (MCM) approach has been that
it encompasses all the "troublesome" components into a single well-controlled unit that
can be tested for total functionality before being put onto a circuit board. It is likely
that, as individual component yields improve, and as the tolerances and losses
achievable with chips auto-placed on high-quality printed circuit boards permit, then the
need for separately testable enclosures will face serious competition on a cost basis.
Another of the issues addressed by the packaging environment is that of heat
dissipation, as an important factor in determining the correct conditions of operation,
and especially for maintaining the correct junction temperature from a reliability
78
viewpoint. Key to the understanding of this facet of IC design is the ability to correctly
model the thermal characteristics of the circuit, including the package and heatsinking
hardware. In spite of the rigorous demands of 'mil-spec' quality requirements, the need to
satisfy commercially-competitive environmental specifications in adverse customer
premises applications (e.g. usable on the wall of a house in either Alaska or Africa) is
driving the packaging technologies to a new domain of cost-effective solutions.
It is fair to say that some of the early GaAs technology proponents oversold their
claims, frequently promising performance advantages based on isolated devices, but
without taking account of manufacturing and operating margins. A consequence of this
enthusiasm was the disappearance of a number of the early GaAs enterprises, and a
major rethink for some of the survivors. As mentioned in section 3.1 above, a realistic
designer is obliged to trade some of the performance margin to compensate for non-
uniformity of devices within a circuit as well as interconnect losses, etc. This
cautionary comment is intended to warn against comparison of technologies in differing
contexts. A research result on an isolated device, or a carefully chosen and "tweaked"
circuit does not necessarily represent what can be achieved on a manufacturable basis. It
is, however, curious to note that, whereas a few years ago Si defenders were pointing
out that GaAs results were only obtained on laboratory specimens, the reverse situation
now prevails, where GaAs IC's are being assailed by highly-tuned Si circuits from
university research groups! A sad fact is that a number of otherwise promising devices
have been culled from the route to development because they were unable to meet their
original expectations, even though in some cases the basic design contained intrinsic
merit.
It is valid to question at this point whether or not commercial viability should be taken
as an objective for selecting future technologies, especially given that some new
technology ideas which are not in themselves commercially viable may nevertheless
spark off a more cost effective version that does provide a better end-result. A
stimulating discussion of the subject by Kroemer [8] highlights the fact that pushing
towards applications solutions and cost effective mass-production tends to limit
creativity and stifle original approaches (such as described in the example of VLSI
digital GaAs development mentioned above).
Nevertheless, my aim here is to consider what type of features of a technology
may enhance its chances of survival to the point of usefulness, in the hopes that fatal
setbacks experienced by other technologies could be averted, or at least anticipated.
Many of these features go beyond the normal technical considerations that might be
addressed in a research paper, and embrace significant aspects such as economic factors,
timing and competition, related supporting technology demands, etc.
description of the costs that anticipates all the potential contributors: materials,
fabrication, yields, testing, packaging, test and package development, reliability
programs, and especially how these costs will evolve as the technology matures.
4.1.11 Does the device have a weak point, fatal flaw or Achilles' heel?
If there is some type of operating condition that is highly unfavorable to the device, it
is better to be "up-front" about it, and to propose design considerations or a
development program to alleviate the impact of the potential weakness.
4.1.12 Does the device need unusual or extra power supplies compared to competitors ?
Certain digital GaAs architectures have found difficulty as "drop-in" replacements for Si,
simply because the logic family needed different power levels. Compliance is better
unless it seriously undermines the performance advantage.
4.1.13 Will the device need to push the edge of some associated technology envelope ?
As in 4.1.6 above, the technology developer may also be faced with developing some
aspect of a related technology on which the new device depends, such as special epitaxy
(e.g.: overgrowth of metal arrays for the PBT), lithography (for nanodimensional
devices), special packaging solutions, cooling technologies for cryogenic devices etc.
4.1.14 Are there environmental factors e.g. toxic chemicals, limited raw materials?
Early allegations suggested that the world supply of metallic gallium would limit the
volume use of GaAs devices; this has not proved to be a problem yet. The toxicity of
As does not emerge as a significant problem either, although some of the metal-organic
chemicals used in epitaxy of certain III-V compounds are being studied to find more
acceptable substitutes.
4.1.15 Are there major technical or commercial risks involved?
This aspect could reflect instances where several technology areas (e.g. 4.1.6, 4.1.7,
4.1.13) could compound to represent a very high level of necessary investment to ensure
success. If this is coupled with an uncertain marketplace and a unique application, then
risk is high.
4.1.16 Are there any potential reliability exposures or sensitivities?
The first heterostructure devices were faced with proving that their atomically thin layers
were not a major liability. Nevertheless, some surprises can occur in a new technology,
e.g.: the "purple plague" in Si ICs, or the H2/gate issue mentioned in 3.4 above.
4.1.17 Does the device require any special operating conditions (e.g. cryogenic cooling)?
Josephson junction devices are a clear example of a technology that had to support a
major "selling" exercise to market the ultra-fast functionality available. Even with a
solution in hand, the unfamiliarity of a major constraint can seriously undermine the
confidence of investors!
4.1.18 Is the operation difficult to explain, or the name hard to remember?
This may seem trivial, but the new idea will have to be "sold" to prospective supporters
of a non-technical background, and if the barrier of comprehension and retention is too
high, then the chance of obtaining the required support may be reduced.
4.1.19 Does the device face comparable responses from competing technologies?
One of the best things to happen to Si over the last 20 years is GaAs. The competition
of GaAs pushed Si technology into new areas that might otherwise have taken longer to
82
emerge. By the same token, the advances of Si sometimes surprised the otherwise
complacent GaAs engineers!
4.1.20 Are there any patents limiting free competitive development"!
Given a marginal choice between two competing technologies, the one which is
constrained by legal limitations may face a battle on a different playing field, at least
until the rights of the party concerned expire, or the limitations are removed.
TABLE 1. Questions applicable to the evaluation of a new technology; for 1-10, 'Yes'
scores 5% and 'No' scores zero; vice versa for 11-20. 'PBT refers to Permeable Base
Transistor [9] and 'QFD' to Quantum Functional Devices [10].
This attempt to quantify the judgments listed above is not intended to be either
exhaustive or exact; for example there is no real assessment of the magnitude of
technical advantage of a proposed technology, nor is there any relative weighting of the
various aspects in generating a "score". The two examples chosen for comparison
against the GaAs FET are the PBT [9] which came into prominence around the 1980s,
and the QFD-type of devices, as described for example in a recent review article [ 10].
The PBT is essentially a vertical field-effect transistor where the gate is made as an
array of electrodes buried within an epitaxial structure. Pinch-off is then between
adjacent fingers of the gate electrode (and therefore more effective than between a gate
and a hi-resistivity interface), and the gate length is determined by the thickness of a
deposited metal layer, and is therefore potentially very short indeed. Excellent
simulation work showed the potential for this device to offer very high frequency
performance, but in spite of the elegant nature of the structure, the practical problems
associated with the overgrowth eventually lead to conclusions that the device would be
overly difficult to manufacture. Variants of the structure (e.g. by implantation of the
gate regions) offer some of the useful features, but sacrifice many of the advantages.
Quantum Functional Devices refers to the various types of III-V (or Si-based)
devices that aim to use tunneling between well-defined states to create I-V characteristics
with regions of NDR (negative differential resistance). With such a multi-state
characteristic, circuit designers could potentially realize certain logic functions or
frequency-doubling operations with a single device where normally several devices
would be needed. The potential savings in device count, interconnect, and chip size
could be significant, however the battle is currently to achieve the requisite functionality
at room temperature and above, and in a manufacturable architecture. Many of the
"answers" in the table are arguable, and their status will certainly improve given the
level of activity on these types of devices.
The discussion of the previous section highlights the fact that a new technology usually
needs a substantial performance advantage if it is to overcome the various other hurdles
en route to industrialization. Typical advances now tend to be factors of 2 or 3 in speed
or power or size. The QFD family may ultimately provide a factor of 5-6 in integration,
and possibly some associated speed/power advantages, but only with a much more
complex technology. Clearly the stage is set for a major change that could offer
improvements of 10-100 times, but that sort of advance does not seem to be on the path
of current semiconductor developments.
New devices are usually selected for further development on the basis of reported
characteristics of peak performance, but only rarely for intrinsic advantages of easier or
cheaper manufacturability. Since the answer to future requirements may not always
mean faster or smaller, it is helpful to reconsider the criteria. Most devices now stem
from evolutions of the basic 3-terminal function, whether bipolar or field-effect in
operation, however an ideal device would have totally flexible coupling from input to
output, with no committed common terminal. Possible new architectures could perform
powerful single-device operations (such as the QFD) and have multiple inputs and
84
outputs, with options for optical interaction or programming of interconnect. Of course
the right technological environment is necessary to encourage such progress.
Some may take exception to the recently-quoted phrase "Real men have Megafabs!"
Apart from the political incorrectness, this suggests that the only solution to cost-
effective manufacturing is scale, and that the only successful technologies are the very
high volume ones. It is easy to show that wafer production facilities operating with
large batches of 12" wafers are inhibiting to the development of small prototype circuits
for low-volume applications. Even with 4" GaAs, the opportunities for, say, university
groups to exercise design ideas in the medium are severely constrained by the needs of
the big bill payers, and flexibility of design rules or rapid turn-around are not generally
available. In the pursuit of lower costs in a given medium, the megafab philosophy
may be creating a monster so large as to find difficulty in maneuvering for survival
when a different, much cheaper technology presents itself.
Neural network technology has many attractive features, but generally involves very
high component counts. Most of the development has therefore tended to focus on
architectures that may be realized within conventional VLSI technology, and relatively
little work has looked at the possibilities of a substantially new technology specific to
neural functions. Features such as variable interconnect, variable thresholds and
weighted responses are key to the adaptability and learning incentives of neural
networks, but are difficult to achieve in hard-wired semiconductor technology.
Nevertheless, the mix of functionality possible within, say, the III-V family (optical,
85
digital, analog) presents some interesting options for e.g. integrated retinal
detection/processing, and some operations such as optimization of digital signal
recovery from optical signal streams might be more efficiently done using a neural
approach than in the conventional manner. It seems unlikely however that such
architectures will really "take off until they are given a more sympathetic medium in
which to operate. When that happens, the learning derived in the current technology will
be able to bloom.
Perhaps a candidate for the necessary material for neural architectures will be
polymer electronics. Although the mobilities of conjugated polymers are substantially
less than we use in semiconductors, the cost is potentially much lower, the volume
usage more efficient, and the potential for applications more flexible (in the true sense
of the word!). It is in this context that one should imagine how the semiconductor
industry would respond to competition from a medium that was so low cost that
volume became no object, and could be conformally applied to the walls/support
structures of any mechanical equipment. There is also potential for an integrated
marriage between the technologies whereby slower operating layers of e.g. polymer
memories/neural routing networks could be deposited onto conventional higher speed
circuits to offer hybrid functionality and intelligent interconnect.
An overriding theme of the discussion in this paper has been towards application-
oriented development. It is tempting to say that the applications must therefore be
clearly identified up front, but history shows that many technologies have to be
developed and matured before their best applications are found. This suggests that
pursuing the trend to cheaper functionality for larger volume applications (because that
pays the bills) may eclipse some valuable new initiatives. Perhaps by focusing further
out on more radically different technologies, we can create some extra flexibility to
distract developers from the limiting trends currently underway. Key features for future
emphasis should be:
• Anticipate the "end-of-the-road" for miniaturization of current devices.
• Avoid the technology development being dominated by "megafabs".
• Pursue the development of small less-precise devices and statistical architectures.
• Encourage true 3-dimensional device and interconnect topologies.
• Develop some cross-disciplinary bridges to e.g. polymer electronics to enhance the
intermediate benefits and lead into the next generation.
• Be aware of the advantages of an easily-manufacturable technology.
• In taking a new technology through to industrialization, expend more up-front
effort on the surrounding issues to smooth the transition.
7. References
1. Jay, P.R. et al. (1991) IEEE GaAs IC Symposium: Short Course Handbook.
2. Peterson, V. (1988) Applications of GaAs IC's in instruments, IN Proc.
10th IEEE GaAs IC Symposium, Nashville, TN, pp. 191-4.
86
3. Malone, H. R. et al. (1991) High volume GaAs MMIC applications, in
Proc. 13th IEEE GaAs IC Symposium, Monterey, CA, pp. 135-138.
4. Thomas, R. N. (1980) Large-diameter, undoped S-I GaAs for high mobility direct
ion-implanted FET technology, in GJ. Rees (ed.), Proc. 1st International Conf.
on Semi-Insulating III-V Materials, Nottingham, 1980, Shiva Publishing, UK,
pp. 76-82.
5. Shannon, L. C. (1990) Increased yield of microwave devices due to subsurface
damage reduction in SI GaAs wafers, in A.G. Milnes and C.J. Miner (eds.),
Proc. 6th International Conf. on Semi-Insulating III-V Materials, Toronto, 1990,
Inst. of Physics Publishing, Bristol, UK, pp. 359-366.
6. Skinner, R. D. (1991) What GaAs chips should cost, in Proc 13th IEEE
GaAs IC Symposium, Monterey, CA, pp. 273-276.
7. Hu, W. W. et al. (1994) Reliability of GaAs p-HEMT under hydrogen containing
atmosphere, in Proc. 16th IEEE GaAs IC Symposium, Philadelphia, PA,
pp. 247-250.
8. Kroemer, H. (1994) Devices for the Future: a peek into the next century, in
Ext. Abstracts of 1994 International Conf. on Solid State Devices & Materials,
Yokohama, pp. 397-399.
9. Nichols, K. B. et al. (1985) Fabrication and performance of GaAs permeable base
transistors, in Proc. IEEE/Cornell Conf on Advanced Concepts in High Speed
Semiconductor Devices & Circuits, Cornell, pp. 61-71.
10. Goronkin, H. et al. (1994) Progress in quantum functional devices to
overcome barriers to ULSI scaling, in Proc. 16th IEEE GaAs IC Symposium,
Philadelphia, PA, pp. 9-12.
COMMENTS ON THE NATIONAL TECHNOLOGY ROADMAP FOR
SEMICONDUCTORS
JAMES F. FREEDMAN
Semiconductor Research Corporation
79 Alexander Dr., Bldg. 4401, Suite 300
Research Triangle Park, NC 27709
Abstract
The SIA National Technology Roadmap for Semiconductors (NTRS) [1] represents a
coordinated effort by leading technologists from all sectors of the U.S. semiconductor
infrastructure. Led by industry and coordinated by the Semiconductor Research
Corporation (SRC) and SEMATECH, this effort involves industry, academia and
government agencies in defining a unified description of the semiconductor technology
requirements for ensuring advancements in the performance of integrated circuits. The
NTRS provides a fifteen-year horizon, extending through the year 2010 and covering an
anticipated six generations of future product and technology needs.
1. NTRS Methodology
Building on the experience gained from the first roadmap [2, 3], the current revision
uses a structured approach to assure unification. A Roadmap Coordinating Group was
established to define the overall strategy and a standardized template. This resulted in the
creation of eight Technology Working Groups (TWG's). Each TWG represents a
technology area critical to semiconductor product development and manufacturing
operations and is assigned the responsibility of identifying the current technology
status, the roadmap of technology needs, potential solutions, other dependencies and
potential paradigm shifts. The eight TWG's (technology areas) are: 1) Design and Test;
2) Process Integration: Devices and Structures; 3) Environment, Safety and Health; 4)
Lithography; 5) Interconnect; 6) Materials and Bulk Processes; 7) Assembly and
Packaging; and 8) Factory Integration.
Since each TWG has a limited number of members, an open workshop was held to
solicit a broad critique of each TWG roadmap prior to publication and release. There was
nation-wide representation and the open discussion resulted in a consensus document.
3. NTRS Framework
4. Analysis
Although technology scaling has been the driving force for technology advancement, it
is the improved cost per function that has generated the market growth and revenues that
89
have fueled the research and development (R&D) costs. For example, of the 4X
improvement in DRAM capacity per generation, 2X has been attained from lithography
(1.4X per linear dimension), 1.5X from chip size increases driven by improved
manufacturing, and the remaining 1.3X from innovation related to cell size reduction
and other technological and architectural advances. Although it is recognized that
continuing improvements in lithography are required to meet the OTCC cadence, these
alone are inadequate to continue the 30% cost-per-function improvement that has fueled
the global growth of the industry.
Consider the following simplified manufacturing cost-per-transistor analysis:
Therefore,
Cb = Cw/(NcxNb) (2)
Where
Cb = bit cost Nc = no. of chips on a wafer
Nb = no. of bits on a chip
Now,
Nc = K2 D2/(Nb x L2 x PE) (3)
Where
K2 = constant L = resolution level (lithographic)
D = wafer diameter PE = packaging efficiencies
Thus,
Cb = K (FC/(FT x Y)) x L2/D2 X PE (4)
Equation (4) demonstrates the importance of the factory (including the use of large
diameter wafers) in addition to lithography and packaging (or design) efficiency.
Factory productivity assessments [5] from 1970-1990 show that wafer starts per
month increased from 10,000 wafers/mo. to 40,000 wafers/mo., yield increased from
30% to the 80-90% range, and wafer diameter increased from 75 mm to 200 mm, which
is the norm in modern factories. The result was a total throughput improvement of 65X
when the total factory cost increased by only 10-15X, resulting in a real capital cost
decrease. The fifteen-year period covered by the NTRS indicates a potential new (and
negative) factory paradigm, where there will be increased capital costs per silicon unit.
Clearly, the yield factor has been almost completely exploited. Factory productivity
demands the movement to not only 300 mm wafers but also to 400 mm wafers mid-
way through the projection period. With the larger wafers and the current equipment
utilization, it is questionable whether a further increase in the volume of wafers per
month will have any beneficial effects, especially when nanometric shrink technology is
driving the need for unit wafer, in-situ processing.
90
5. Challenges
Considerations such as these were factored into the NTRS, resulting in the
acknowledgment that the semiconductor industry faces new challenges as it moves to
production of feature sizes that are less than 150 nm.
Clearly, the first challenge is to develop the cost-competitive lithographic
processes that will allow exposures of sub-150 nm lines. It is generally recognized that
optical (deep-UV) lithography will be extended and used down to 0.18 (im. However,
this extension requires complex enhancements with a large development expense —
both for the tool and for the process. Critical dimension control and overlay are
becoming crucial, placing an increasing demand on metrology tools and masks. The
challenge is to contain the manufacturing cost associated with this increased
complexity. In addition, there is no clear manufacturing solution for dimensions below
100 nm.
Furthermore, the NTRS identified four challenges that span the entire spectrum of
technology and require a major effort to resolve. Termed the "Grand Challenges," they
are:
4. Technology funding: The cost of new factory capital and the escalating cost of
R&D can only be funded out of current profits. The pressure has already driven the
industry to establish cost-sharing consortia like SRC and SEMATECH and to establish
individual partnerships and other relationships. The trend towards common research
funding will and must continue as solutions are defined to address the issues discussed
91
above. The reality is that the available funds will never be adequate to address all the
issues raised as trie industry moves to nanotechnology. The challenge is to implement a
funding strategy that covers all critical needs and to demonstrate the viability of new
concepts prior to funding the development of commercializable solutions.
6. Summary
By focusing on needs rather than solutions, the National Technology Roadmap for
Semiconductors provides a broad description of the technology challenges requiring
resolution in order to extend the integrated circuit revolution into the 21st century. It
realistically addresses the major barriers to this extension, recognizing that the cadence
will not continue if these barriers are not removed. The challenges are great, but the
promise is even greater.
7. References
PAUL M. SOLOMON
IBM Research Division
T. J. Watson Research Center
P. O. Box 218
Yorktown Heights, NY 10598
Abstract
The capability of silicon technology has increased and cost of doing computation has
decreased to the point where vastly expanded personal computational facilities become
available to a large class of users. In this environment the energy cost of computation
becomes a critical issue, especially for portable applications, but even in desktop,
household and office environments. The capability of the upcoming technologies to
deliver performance for some future yielded chip will vastly exceed the power budget
allotted to the application desirous of using that performance. The concept of "excess
capacity" is introduced to describe this situation and this excess capacity may be traded
for power in different ways to realize different system solutions. This paper will discuss
the nature of the different kinds of trade-offs, and the classes of system solutions
realized.
The semiconductor industry has seen unprecedented growth over the past three decades,
forming the basis a great social revolution, akin to the industrial revolution of two
hundred and fifty years ago. Since its advent we have seen a million-fold increase in the
scale of integration and in the capability of electronic systems, which have become
increasingly available, through their reduced cost, for use by the average citizen.
The pattern of development of computing systems, since the invention in the
1970's of the single chip microprocessor, has been a hierarchy of large, fast and
expensive mainframe computers (high end) contrasted with small, slow, yet cheap
microcomputers (low end). As technology has progressed, the capabilities at both the
low and high end have increased both in terms of speed and numbers of circuits. The
number of memory circuits has increased much faster than the logic, with slower
memory chips serving fast logic chips. The objective, at the low end, has been to fit an
entire CPU (central processing unit) on one chip, and this has been achieved by reducing
the number of bits being processed at a time, starting at 4b in the early 70's and
progressing to 8b, 16b and 32b as the technology advanced, compared to 64b in a
93
S. Luryi et cd. (eds.), Future Trends in Microelectronics, 93-109.
© 1996 IBM. Printed in the Netherlands.
94
typical mainframe. Now with 64b CPU chips becoming available, we see a convergence
between the low and high end in terms of single CPU capability.
As a result of this phenomenal growth, the average user in his office or home now
has personal access to computational power that was once reserved for mainframes, and
a plethora of applications has developed to exploit this capability. Fortunately the
electrical power per unit of computation has also decreased dramatically, permitting
these computers to operate at power levels consistent with home or office use. Indeed,
the power has been reduced so dramatically that an entirely new class of computers, the
portable computer, has arisen. This accelerates the trend toward more computation at
ever reduced power levels and leads to the first postulate for the present critique: The
power budget for the most important computer applications will decrease over time.
The assumption here is that home, office, and portable applications will be the most
important (in terms of market share), and forces driving the power budget down are
primarily increased battery life for portable applications, and energy savings for home
and office computers. Indeed, it has been estimated [1] that today, computers consume =
5% of the nation's electrical energy!
Historically, power has not been an important constraint for low-end processor
design within the boundaries imposed by complementary metal oxide silicon (CMOS)
design practice. Compared to bipolar and n-channel FET technologies, which consume
power even when idle, CMOS offered such large power savings for systems running at
low clock rates that technology was able to advance, on a fairly conservative scaling
path, limited by speed and area constraints, while keeping the power dissipation of the
CPU chip fairly small. The only concern related to power was the ability to cool the
chip, since most of the power consumption of the desk-top computer lay in the display.
Since its introduction, CMOS technology has been able to proceed toward higher
densities, without a major collision with power constraints because of a combination of
circuit innovations and the application of scaling principles. Indeed, according to the
scaling principles laid out by Dennard et al. [2] more than two decades ago, FET
integrated circuits can be scaled to smaller dimensions, higher densities and higher
speeds without increasing power density. This is easily derived from the dependence,
P = NcktCV2f (1)
of the power on the number of circuits switching, capacitance per circuit and voltage
being switched, and switching frequency, where capacitance, voltage and logic delay
(inverse of the maximum switching frequency) scale linearly with dimension. For
several generations the voltage had remained constant at 5 V, to maintain compatibility
with the old TTL logic interface circuitry, and power had been kept in check by
combination of circuit, layout, and architectural innovations.
Further progress in scaling has demanded a return to scaling the power supply
voltages, hence the new 3.3 V standard, as well as 2.5 V and 1.8 V supply voltages
proposed for the future. At some point voltage scaling becomes harder to do, as we will
discuss further, and the voltages will be constrained to be larger than some minimum
which depends on various system parameters.
At a constant voltage, and with no further innovation in layout, circuit techniques,
and architecture, the power density will increase rapidly with scaling as the inverse
square of the dimension. If past trends are followed, chip size will continue to increase
and the total power per system would increase dramatically for systems at the cutting
edge of speed and integration capabilities of the technology. This leads to the second
95
postulate of this study: The power dissipation of a chip designed for maximum speed
and integration potential of the current technology will increase dramatically with time.
Now one may argue, as is done in Ref. [3], that chip size need not increase and will
actually decrease with time once a full 64b CPU and its cache can be fit onto one chip.
This is difficult to dispute, but flies in the face of the economic incentive to exploit the
new technology to its fullest extent.
The two postulates of our study, as illustrated in Fig. 1, lead to a dramatic
divergence between the power budget for the major applications and the power
dissipation of a chip operating at maximum capacity. This trend is also document in
recent work by Stork [4]. In order to reconcile the two, the processing capacity of the
chip has to be reduced until the power fits within the power budget. We will call the
ratio between the maximum capability of the technology and the actual processing
capacity at a given power budget the excess capacity. Using reasonable trends [4, 5] we
find an excess capacity of between one and three orders of magnitude for desk-top and
portable computer applications for the 0.1 urn technology generation.
Given this excess capacity, the question becomes how to trade it for lower power
consumption in the most efficient manner. This will be the subject of the rest of this
paper.
10°
CHIP POWER
AT MAX.
POWER CAPACITY
BUDGET
10z
o
0.
-2
10
TECHNOLOGY EVOLUTION
Figure 1. Divergence of chip power dissipation from the power budget for typical
applications as a function of technology evolution (time).
Figure 2. (a) Charging of capacitor C via switched inductor, (b) Charging capacitor
via resistive FET.
The circuit of Fig. 2 (a) illustrates how a node may be charged and discharged
dissipationlessly, assuming that the circuit elements were lossless, clearly showing that
the paradigm represented by Eq. 1 is not universal. In this figure the switch represents
an FET, the gate of which is controlled by other adiabatic circuits. A simpler
representation is shown in Fig. 2 (b) where the switched FET is represented by a
resistor, R, and the capacitor C represents the gate plus wiring capacitance of the
circuits being driven. The circuit is fed by a ramped power supply of rise time T. For
this example, the energy dissipation per transition,
U = CV2(RC/T) (2)
A "S.
B ^ **»•'
r /
OKI s s.
CK? /•
CM TN
y-vl
CK4
Olli y
►
time
While this is one of the simplest ways of applying adiabatic concepts to CMOS
logic it is by no means the only way [10]. For instance, one can achieve a reversible
pipeline by using inverse logic functions to implement a reverse pipeline which
adiabatically resets the logic stages in the wake of the forward propagating signals [11].
10r«
ai i io
CLOCK FREQUENCY (MHz)
The property of adiabatic circuits is that for a given Vßß the power is proportional
to the square of the frequency. This was verified by us (see Fig. 4) in the case of a
98
simple adiabatic buffer circuit, by measuring the small temperature rise of the chip
caused by the dissipated power. This means that at a low enough frequency the adiabatic
circuits will dissipate less power than conventional circuits.
Fig. 2 (a) represents a possible on-chip solution for powering adiabatic circuits
requiring a low-loss on-chip inductor [12]. Unfortunately such inductors are impractical
to make on a VLSI chip. Apart from the considerable problems of incorporating micron
sized magnetic components on chip, the Q of a given inductor design is reduced linearly
with dimension so that conventional inductor designs would not work well at micron
sized dimensions. More practical schemes use capacitor networks [13] but these have
limited potential for energy savings.
(a) (b)
s
r1-i"
V
-'"~
SUB
Practical CMOS latch circuits, each in succession approaching that limit more
closely, are shown in Fig. 5 (a-d). The first achieves efficiency by minimizing the
capacitance of the latch and buffering it from the larger load capacitance with an
adiabatic buffer. The second employs a series diode, making the cost of energy erasure =
CVDDVfade, so the energy saved is Vdiode^DD, where Vdiode = (kT/e)ln(Ion/I0ff).
When the 'on/off current ratio is optimized for minimum power dissipation, Vdiode =
99
(kTle)\n{eV[)£)lr\kT) is the diode for ward voltage drop, and r\ (7] < 1) is an activity
factor. A further power saving may be achieved [14] using a cross-coupled latch for data
storage (Fig. 5 (c)). For reasonably fast erasure the last Vj of voltage on the internal
nodes must be erased non-adiabatically leading to a penalty proportional to CVjZ. For
extremely slow erasure the energy cost will ultimately be reduced to several times
(kTlefiC. This same reduced cost can be achieved in a reasonably short time by ramping
down the threshold voltage during latch erasure [15], for instance by changing the
transistor well bias locally as shown in Fig. 5 (d). We can relate the above result to the
more fundamental result of Landauer by expressing it as nkT where n is the number of
electrons stored on the capacitor C. Landauer's limit is therefore approached in this kind
of latch when the FET's are small enough to store just a single electron!
Because partially adiabatic latches can be made so much more energy-efficient than
conventional CMOS, even including a large number of latches into an adiabatic system
can result in substantial power savings. As an extreme case, the scheme of Denker et al.
[16] uses every gate as a latch and realizes a 3:1 power reduction.
Multi-phase
Energy-Recovering
AC Signal
Source
> ,n
1 '
Shift Register 1
n ^ >
Matrix
SwlItch
Data
In =! Block -
Figure 6. Hypothetical adiabatic logic chip, from Frank and Solomon [18].
5. Voltage Reduction
The total power dissipation consists of both dynamic power (CVjyD^f) and the
static power due to the Nckt l0ff^DD 'off current. The two can be related through Eq. 3
and the condition that the maximum switching time tmax be less than the clock period.
Here Cmax is the capacitance of the slowest circuit, as opposed to the average
capacitance C used above. The optimum power supply voltage Vßßopt to minimize
the total power is then:
where Nckt is the total number of circuits, Nsw is the number of circuits being
switched, Fj is the average fan-in, and Fjiinax is the fan-in associated with the slowest
circuit.
Tolerances become extremely important at low supply voltages. Of the tolerances
we will consider only the Vj tolerance (assumed equal for p and n-channel FET's for the
sake of simplicity). Other tolerances, such as those of the power supplies, are
important, but not strategically so, since it is quite feasible to regulate the power
supply on-chip, even at multiple points. The simplest way to include the Vj variations
is increase the power supply voltage by them:
The reason for adding both AVj+ and AVj~ is that AVj' subtracts from Vj,
requiring a higher nominal AVj, while AVj+ raises Vj above nominal, requiring an
even higher power supply voltage to achieve the speed objectives.
0.6
0.4
/T* nMOS
S. 0.2 -1.8V(Vds)
o
pMOS
-Til -t
-0.6
0 .1 .2 .3 .4 .5 .6 .7 .8 .9
Effective Channel Length (/um)
Figure 7. Threshold voltage vs. gate length for 0.15 (Am CMOS, from Davari et al. [5].
102
Threshold voltage variation is due to many different factors, including doping
profiles, work function, well bias and temperature, etc., but we will consider only
channel length since this is the most difficult to control. Threshold voltage varies with
channel length because of the interpenetration of source and drain fields into the channel
region and their screening effect on threshold control implants. Typical Vj roll-off
curves are shown in Fig. 7 for a 0.1 |J,m technology. In a sense, these curves define the
technology, both in terms of channel length control and in terms of vertical profile, so
that for a given state of the technology there is a clear trade-off between channel length
and Vj1 control. In order to get the tighter Vj control necessary for low voltage
operation, the channel length for a given technology would have to be increased.
The parameters in Eq. 4 reflect many possible choices in system architecture,
design methodology, and layout. Such parameters are the ratio of switching to total
circuits, the ratio of maximum switching time to clock period, and the ratio of average
to maximum capacitance. A range of choices is covered in Table 1, the first row
representing a large general purpose chip, while the second a specialized function
optimized for low voltage. The range of VDD.min (0-2—1 V) represented by these
choices is considerable and is an indication that future chips may well depart from the
single, standard power supply voltage that has been used up to now.
Portable systems run off fixed voltage batteries of course, and lithium batteries, at
~3 V, are increasingly being used because of their high storage capacity. This voltage
must be regulated down to whatever power supply voltage is appropriate. The losses in
the regulator will introduce extra factors into the optimization of VJJQ, resulting in
somewhat higher values of Vßßmin.
mkT c /'max Fi F,
1
Itmax AVT+ AVf 'DD.min
e N c
/mV\ (mV) (mV) (V)
V Dec )
90 104 100 0.05 3 4 100 150 1.1
60 10 3 0.5 2 3 25 50 0.28
At a constant Vjit can be shown [11, 17] that power dissipation is minimized at
Vf/Vßß ~ 3.4. This is when the effect of the resistance increase of R due to reduced
gate voltage just balances that of the reduced CV2.
When the Vj-ZVßß voltage ratio is held constant, adiabatic circuits yield the same
square law dependence (CVßj)2RC/T) of switching energy on voltage as nonadiabatic
circuits, assuming that T and R are both inversely proportional to I/VDQ. TO find the
minimum power supply voltage for adiabatic circuits we revert to our sub-threshold
model and optimize the power supply voltage using similar procedures as before:
103
N
T/ _ 2mkT ,„ T cu ( T VI ,n
(6)
VDD,opt—-T-^l-^^W) J
where /? is effective average channel resistance of the 'on' FET's at a source voltage of
half the power supply voltage. This resistance can be adjusted by choosing an
appropriate threshold voltage. Now Eq. 6 resembles Eq. 4 the main differences being the
factor of 2 outside and the RC/T replacing the ftmax inside the logarithm. The resulting
optimal power supply voltage is therefore larger by approximately a factor of two than
in the dissipative case. Tolerances still have to be added, as in Eq. 5. One might argue
that adiabatic circuits sense average rather than extreme values because power
dissipation, rather than delay, is the issue. This is true in principle, yet in the sub-
threshold regime extreme values tend to dominate due to the exponential dependencies.
The Vßß limits we have derived so far apply only to simple static logic circuits which
are very robust in terms of noise margin. Other circuits generally require higher supply
voltages, especially circuits involving pass gates. Such circuits include the ubiquitous
static memory cell.
Pass transistors require approximately an extra Vj of supply voltage because two
G/S voltage drops are in series. In the sub-threshold regime we can derive the extra
voltage by requiring that the delay penalty of the pass transistor and the following logic
gate be equal to the delay of the heavily loaded circuit discussed above, resulting in:
where V/)D" is the value of Vpß for the system with no pass transistors. Inserting
typical values, AVßß - Vßß< - 0.25 V. The penalty is therefore much larger under
conditions that require a larger Vpp in the first place and approaches doubling of the
original Vßß .
In a static memory cell, as shown in Fig. 8, the bit line access transistors (n-
channel) act as pass transistors; however, the cell is read and written differentially so
that at least one of these transistors will always have its source close to ground
potential during the entire operation. For low power operation it is feasible to design
the cell and sense amplifier such that the access time does not depend critically on the
access transistor on the ungrounded side. For the READ this need not depart much from
some modern designs [23] in which sensitive sense amplifiers can detect small
differential signals. The WRITE operations will require some changes to the standard
design practice. While pull-down via conventional sense amplifier-decoder is
satisfactory, the pull-up via the pass transistors will be very weak, so that the p-channel
transistor in the cell would have to be strong enough to provide this function within a
WRITE access time.
104
WORD
LINE
Figure 8. Static memory cell with separate array power supplies.
The above discussion illustrates how the voltage penalty due to the pass
transistors in static memory can be circumvented by means of appropriate design
technique. Some common practices, such as the use of area-saving very high impedance
pull-up devices, could prove unsuitable for low-voltage design.
Pass transistors for LOGIC are not strictly necessary. For instance any
combinatorial logic function can be built from simple NAND gates. Therefore the CPU
and static RAM could in theory be built without incurring the voltage penalty of pass
transistors. The lesson to be learned, however, is that an additional trade-off in low-
power (voltage) design is the necessity to abandon certain circuit techniques, which are
often used to obtain higher speed at a conventional Vߣ>. This becomes just another part
of the speed-power-density trade-off.
8. Sleep Modes
An increasingly used low power design technique is the power down of large blocks of
circuitry that are not being used for a particular calculation. This eliminates the DC
leakage current associated with that circuitry, but of course requires that the power-down
circuitry itself have low leakage current. We see the utility of this technique from Eq. 4,
where the threshold voltage could be reduced (to increase Ion) and the power supply
voltage reduced proportionally to the ratio of Nsw to Nc^t, where now N^t is the
number of (low Vj) transistors in the non-sleep mode, and the transistors in the power-
down circuitry have high Vj to reduce leakage.
One can easily see how such a technique may be applied to logic, but for memory
the stored information may be lost if the array is powered down. On the other hand, the
need for a sleep-mode is especially acute for static memory which may comprise the
bulk of the circuitry on a modern CPU chip.
105
Low standby power static memory requires transistors in the cell which have high
threshold voltages; and this conflicts with the goal of low-voltage operation. This
dilemma may be resolved if the voltage powering the array be higher, and the internal
voltage swings be larger than those of the peripheral circuits. This is possible because
the static memory cell has internal gain. Keeping the voltage swing on the peripheral
circuits low, including bit lines, is beneficial because these circuits consume most of
the dynamic power of the array. In order to complete this picture the voltage swing on
the word lines must also be kept low, but this conflicts with leakage requirements as
seen from Eq. 3. To resolve this dilemma we must introduce two modes of operation of
the array, ACTIVE and SLEEP mode. During active mode the array operates with
normal voltage swings on the gates of the access transistors and the leakage current is
tolerated. During SLEEP mode this leakage path is shut off by applying an appropriate
bias. An alternative, but technologically more difficult, method would be to increase the
threshold voltages during the SLEEP mode.
This hierarchy of modes is particularly suited to the already existing hierarchy in
the memory organization where a small, fast cache is served by a much larger, slower
cache, parts of which may be in a SLEEP mode.
9. Comparison of Trade-Offs
As we have seen, a simple reduction of clock rate does not reduce the energy required for
a given computation. As noted by Horowitz [24], this energy is reduced when other
system parameters are changed in the course of the optimization toward low power
operation. These parameters may be voltage (which we have discussed at length), device
size, circuit design, etc. Horowitz noticed that the energy time product (Ut) was
approximately constant over a range of many parameters. Table 2 summarizes the
behavior of FET circuits for some of these variations. Included in this list are adiabatic
circuits, where the clock frequency is the variable.
It is seen that as one varies the parameter of interest that at one extreme (high
voltage, large area) the system exhibits little flexibility in speed with respect to changes
in power dissipation, while at the other extreme there is little trade-off of energy for
speed. At intermediate values, and for adiabatic circuits, there is considerable flexibility
in trading speed for energy dissipation, and a region exists which has an approximately
constant Ut product. The constant Ut region (also called action), while having no
106
particular theoretical underpinnings, is nevertheless a useful heuristic with which to
characterize the effects of design changes.
For the different techniques listed in Table 2 the total energy savings realized in
traversing the trade-off region are very different: less than one order of magnitude for the
case of size variations, about two orders of magnitude for voltage, and potentially many
orders of magnitude for adiabatic computation.
In the trade-off region discussed above, energy may be traded for delay, so that by
slowing down the computation its total energy cost is reduced. This may also be
achieved by partitioning the same computation among several processors, if the
problem so lends itself, resulting in the same or even greater computational throughput
but at reduced energy cost.
At a constant Ut product, and a desired total throughput, the total energy required
for a given computation will be inversely proportional to the number of processors in
the system. As discussed in the introduction, this choice will become increasingly
available for future technologies where many complete processors could fit onto a single
chip. As an example, let us consider the technology choices outlined by Davari et al.
[5]. At the 0.1 \im level this technology will be able to support a RISC processor on
an area of about 18 mm2, so that a 2 cm2 chip would support about 10 such
processors. For this example, parallelism reduces the chip power by a factor of 10 while
maintaining the same throughput.
With this approach, the granularity of the larger system on a chip will be that of a
single CPU and its associated cache. Other large CPU-sized special purpose functional
blocks will be included in this mix. An advantage of this partitioning, apart from the
benefits discussed above, is that for work loads not requiring all of the CPU's, the
unused CPU's can be put into the SLEEP mode with all of the attendant power savings.
In the introduction we presented the thesis that the driving force behind the quest for low
power was the advance of technology. A technology road map into the next century is
presented, for instance, by Davari et al. [5]. In this section we will consider these
options in more detail.
The primary driving force of the technology is the increase in density which, as we
have discussed previously, will allow for multiple processors and large static RAM's on
chip. To achieve the goals for low power, the voltages will be driven down toward their
lower limit and SLEEP mode will be used extensively. DC-DC converters will find
common usage on-chip to regulate the battery or external power supply voltage down to
the chip power supplies.
Device scaling will be pushed to the limit with MOSFET's having effective
channel lengths of less than 0.1 (im. At such short channel lengths the carrier velocities
in the channel would be saturated at voltages low as 0.5 V, so that reducing voltages
toward their minimum would not necessitate a severe delay penalty. In a constant
voltage scaling scenario, adiabatic circuits fare better than conventional circuits because
the channel voltage is always small and velocities do not saturate. Anything done to
107
improve the channel mobility, such as the inclusion of a germanium alloy, will
therefore help adiabatic more than non-adiabatic circuits.
The crucial issue for scaling and for voltage reduction is Vj. Further progress on
the latter front is very difficult to achieve due to the technological difficulty of
controlling the dopant profile, but there are no fundamental limitations to Vj control
down to dimensions of at least 0.05 |J.m effective channel length [25].
Wire resistance will be an important limiting factor to achieving very high performance
in future high speed circuits. This trend will reinforce the move toward parallelism,
since the slower individual processors will be much less limited by the wiring delay.
The scaling potential of adiabatic chips is limited by wire resistance. The power
dissipated in the wires, Pw is proportional to/5/2/X2 for a constant-sized chip [17],
where X is the scaling parameter. This relationship reverses the otherwise favorable
trend for adiabatic circuits compared to conventional circuits.
As an example we estimate the magnitude of this effect using the same scaled
version of Davari's processor [5] as before, where the capacitance C of the processor is
derived from his data on power dissipation to be 5 nF. Assuming an AC power
distribution bus of 10 |J,m thickness (the skin depth at 130 MHz) fed in from the side,
the power dissipated in the wires is 80 mW for the single processor. If a larger chip
were used and 16 processors were to operate simultaneously at 1/16 the frequency (and
with a 40 |Xm thick power distribution bus), the power dissipation in the bus would be
reduced to 1 mW. Note that the thick power bus could be situated off chip and the AC
voltage transmitted to the chip via an aerial array.
The above example illustrates that the wire resistance limit to the scaling of
adiabatic circuits is several generations beyond today's technology, but will be
encountered before the scaling of FET technology has run its course.
13. Conclusions
13. References
STEVE NELSON
Steve Nelson and Associates
6706 N. Lakeshore Drive
Chippewa Falls, WI 54529
1. Introduction
In 1993 the Semiconductor Industry Association (SIA) projected the future capabilities
of commodity integrated circuit devices into the next century. Table 1 charts line width
reductions of a factor of about seven during a sixteen year period.
Chip size (mm2) - - logic / microproc 250 400 600 800 1000 1250
-DRAM 132 200 320 500 700 1000
Wafer diameter (mm) 200 200 200-400 200-400 200-400 200-400
Defect density (defects 1 cm2) 0.1 0.05 0.03 0.01 0.004 0.002
No. of interconnect levels - logic 3 4-5 5 5-6 6 6-7
Maximum power (W / die) - high pert. 10 15 30 40 40-120 40-200
- portable 3 4 4 4 4 4
Power supply voltage (V) - desktop 5 3.3 2.2 22 1.5 1.5
- portable 3.3 2.2 2.2 1.5 1.5 1.5
Number of l/Os 500 750 1500 2000 3500 5000
Performance (MHz) - off chip 60 100 175 250 350 500
-on chip 120 200 350 500 700 1000
If the area is to double (from 1995 to 2001), the IC perimeter will only increase
by a factor of about 1.4. Yet the number of projected I/O pads is projected to increase by
a factor of 2.66. I/O requirements will drive two dimensional arrays of I/O pads on IC's
which can only be connected with technology that closely couples the substrates to the
113
IC's. This has the effect of more tightly coupling the substrate wiring to the device
wiring. In fact, pad-to-pad connections on the same IC will even be sometimes routed
the board-level substrate to provide configurability, to solve a tough on-chip routing
problem, or even to reduce propagation delay. The packaging technology which allows
this sandwich of silicon and substrate can be the base of an even more aggressive three
dimensional packaging system. The motivation to connect large numbers of integrated
circuits to build a traditional high performing single CPU (as used in the CRAY-1) will
be history. This will, however, be replaced by new motivation to effectively connect
large numbers of parallel CPU resources and to tightly couple coherent cache memories
into the architecture.
Pad Array IC
i
Technology
rssssssssssssssssssMrssssssssssssMrssfssssrj'sssssssjwsssssssssfssssi
1 Thermal Management
Cache is introduced into system architectures to reduce latencies to main memory. But
adding additional levels of hierarchy to the memory subsystem increases system
complexity dramatically. Virtual memory address translation and multiprocessor
configurations provide additional challenges which must be met. It is fair to say that no
system has been built to date that can provide a simple cache coherency model, high
performance, and scalability to very large numbers of processors. At least one of these
objectives is invariably compromised.
Associativity describes the degree of flexibility the system has in placing cache blocks
(lines) from main memory into the cache. Direct-mapped is the easiest to implement
but provides for only one "slot" where a main memory block can be moved to the
cache. This generates a higher possibility of conflict among memory blocks and can
cause a debilitating characteristic called "thrashing," an effect where blocks move
inefficiently back and forth between cache and main memory. At the other extreme, a
fully associative organization provides a cache with slots that can receive any memory
block. While the performance of a fully associative cache is almost always superior, it
114
is typically prohibitive in cost. This cost is primarily borne in the address translation
and cache tag identification circuitry. Quickly locating a cache block requires a highly
parallel translation and a search usually using content-addressable memories (CAM).
Large scale content-addressable memories are very expensive to build due to the extra
active devices required for each memory cell and added wiring complexity.
The n-way set associative cache is a compromise solution, where n is the number
of cache block locations that any main memory block copy can occupy. This greatly
restricts the size of the CAM. A classic study of cache miss rates (the frequency of CPU
references which do not find the data already in cache) versus cache size and associativity
was made by Hill. This study showed that as the degree of associativity increases, the
miss rate decreases. In addition, for large caches with higher levels of associativity,
improvements in the miss rate become vanishingly small. Looking at such cache miss
data, a computer architecture designer would probably conclude that eight-way
associativity is enough. However, the cache miss table data is badly skewed away from
typical numerical workloads which exhibit a very different address pattern than the
mixed workloads used in the study. Due to underlying array data structures these codes
generate frequent non-unit stride memory address reference patterns as shown in the two
examples of Fig. 2.
0
1 »
2
3
4
Array A
Array B
n
Array X
Referencing Xi, Xi+n, Xi+2n, etc. Referencing B[ Ai ]
Non-unit stride single-word references have the effect of greatly increasing the
relative cost of a large block (line) cache load. If only 8 bytes are required but an entire
line, of perhaps 64 bytes, is moved the memory port is operating at only 1/8 efficiency
for this pattern. This is the primary reason why vector architectures have traditionally
eschewed cache memories.
How might three dimensional packaging address this problem? Numerical codes on
scalar architectures would do much better if the cache were organized with very small
line sizes (probably only one word — 8 bytes — in length). Additionally, the cache
115
should be very large. If it cannot be made fully associative, it should at least be "many-
way" set associative to avoid serious contention and excessive miss rates. Such a cache
subsystem will contain a very high degree of wiring complexity to adequately drive the
many sets and to implement the large content addressable memory.
CPU (slow)
DRAM
100%
1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992
Figure 3. Historical DRAM and CPU performance (from Patterson, D. and Hennessy, J.
(1990)Computer Architecture: a Quantitative Approach , Morgan Kaufman, San
Mateo).
256 bits in 35 nsec => 7.31 Gbits / sec Leading production SRAMs
(Motorola BurstRAMS)
The internal aggregate data rate of the memory array sense amps is much higher than
that transmitted through typical 1-bit or 4-bit wide DRAM packages. Figure 4
demonstrates how much room there is for improvement if this resource could be tapped.
But bringing 128 or 256 bits off of the DRAM would easily overwhelm a traditional
packaging environment. Only area die interconnect and three dimensional stacking of
devices are likely to succeed.
Rambus is a company that has been successfully licensing a 500 MByte/sec DRAM I/O
technology. The Rambus DRAM channel requires high-speed drivers and receivers both
within the DRAM's and the memory subsystem ASIC's or the microprocessor.
The Rambus technical guide states, "The physical length of one Rambus Channel
is presently limited (to approximately 10 cm) by the 2 nanosecond propagation time of
signals from one end to the other. This length can accommodate up to 32 RDRAMs or
up to ten memory RModules, or combinations of the two. Ten memory RModules hold
up to 320 DRAM'S, giving a total of 160 MBytes of memory capacity using 4 Mbit
RDRAMs." (16 Mbit devices will be offered soon to allow 640 MBytes of storage.)
This total package is roughly 16 in2 (4x4x1 inch) and contains about 30% silicon.
It would appear that a more aggressive stacking technology (but with improved thermal
management support) would allow an increase in bit density of about three and still
obey the 10 cm rule. The other possibility is a potential increase in the clock rate with
a smaller configuration by using improved packaging technology.
Ramtron has introduced a merged device that provides cache buffering of slower but
more dense RAM cells. However the actual sizes of the individual cache and RAM
components are quite modest. OKI Electric has just reported on a 32-Bank 256 Mbit
DRAM with cache and tag. A three dimensional packaging technology combining many
broadside-I/O DRAM'S and SRAM's acting as cache could extend the goals of the
Ramtron and OKI devices for single processor systems. Then, adding coherency support
between multiple modules would add significant performance to symmetrical multi-
processor nodes. Again this would require high wiring density and close spacing.
One way to improve parallelism in a vector processor is to add multiple vector pipes.
The CRAY C90 uses dual pipes as shown in Fig. 5. Simultaneously, two operands
pairs are extracted from the vector registers and routed through dual functional units,
such as the multiply units in the figure. More pipes, more parallelism. (It is important
to realize that the benefits for short vectors will be less pronounced, due to vector start-
up time.) But for large matrices and long vectors multiple pipes can greatly improve
performance. The price to pay, however, will be a large increase in wiring density. This
is one example of expanding a state-of-the-art high-performance-processor in a way that
necessarily requires multiple integrated circuits. Very high density wiring to and from
118
the memory subsystem will be needed to adequately feed the large number of vector
operations.
For an eight- or sixteen-pipe design the signal count becomes very high indeed.
And every data path is latency sensitive so each must be as short as possible. Only a
three dimensional technology could meet all of these requirements.
Ri = Aj*Bj
B c
A 1 V 11 AA
Single Pipe
\ Ri = Ai*B|
Dual Pipe
Rj+1=Ai+1*Bi+1
Advances in integrated circuit density as applied to DRAM technology will very soon
provide a watershed opportunity for large system design. This will occur at about the
time DRAM densities achieve 256 Mbit integration levels. Then a single chip node will
be possible which contains 8 or 16 MBytes of storage, sufficient cache memory, a
complete 64-bit microprocessor with multiprocessor synchronization support, and a
three dimensional router (Fig. 6). Due to the design point of the DRAM and power
considerations, this will probably not be the fastest processor of its generation, but for
certain applications performance can be made up by using very large numbers of
processors. Consider a large system with more than 10,000 64-bit processing elements.
119
8 to 16 MBytes of storage
32 K Bytes of cache
32 or 64 bit CPU
Router with > 100 MBytes
of bandwidth over each
Synchronization Support channel
Router
Using only modestly aggressive I/O driver circuitry, very respectable I/O rates over
each router path are possible: 100 MBytes/sec on each of the twelve paths would be a
reasonable starting design point. The three-dimensional router would allow a folded
torus topology.
This FG-MIMD system would introduce some important tradeoffs that affect the
selection of application suites which would run efficiently:
i) Because the processors would not be the fastest of the generation, a larger number
would be required to achieve a given peak performance level. This means that
communication latencies would be longer.
ii) In a single chip per node implementation the amount of memory per processor
would be modest.
Jack Dennis from MIT described the data flow concept in a 1979 paper at the First
International Conference on Distributed Computing Systems entitled "The varieties of
data flow computers". He said: "Fundamentally, the data flow concept is a different way
of looking at instruction execution in machine level programs — an alternative to the
von Neumann idea of sequential instruction execution. In a data flow computer, an
instruction is ready for execution when its operands have arrived - there is no concept of
'control flow', and data flow computers do not have program location counters. A
consequence of data-activated instruction execution is that many instructions of a data
120
flow program may be available for execution at once. Thus highly concurrent
computation is a natural accompaniment of the data flow idea".
Figure 7 graphically shows the machine code for a simple arithmetic instruction.
With many values being created and presented as input tokens that "fire" when the data
arrives, there is a huge opportunity for concurrency and, therefore, high performance.
ADD
A
B
ADD MUL
AVERAGE
0.87
ADD
C
D
Figure 7. Data flow machine code for an arithmetic expression.
The original data flow model proposed by Jack Dennis at MIT is now known as
the static data flow model. An alternative, the dynamic data flow model was proposed by
Arvind, a collaborator with Dennis in early data flow research. The detection of
matching tokens in data flow computer requires very large amounts of content
addressable memory. In addition there is a problem of allocating resources when a code-
block is mapped to a processor mapping unit. If this resource becomes overcommitted,
the processor can deadlock. As already described, content-addressable memory requires
more wiring than directly addressable memory. A planar array of silicon devices
embedded within a three dimensional wiring array to transfer and buffer the tokens to be
matched would be an important enabler for the data flow architecture, which has never
delivered on its promise because of serious implementation obstacles.
The wiring paths to be optimized involving VLIW cache are, of course, extremely
latency-sensitive. A compact three-dimensional packaging approach would reduce wiring
lengths and allow increased fan-out for synchronization and coherency signaling.
integer integer floating point floating point memory ref memory ref branch
D
"A VLIW instruction might include two integer operations, two
floating-point operations, two memory reference, and a
branch. An instruction would have a set of fields for each
functional unit-perhaps 16 to 24 bits per unit, yielding an
instruction length of between 112 and 168 bits."
Figure 8. Very Long Instruction Word (VLIW) architectures (from Patterson, D. and
Hennessy, J. (1990) Computer Architecture: a Quantitative Approach,
Morgan Kaufman, San Mateo).
"A Subnanosecond Clock Josephson 4-bit Processor," Kotani, Imamura, Hasuo, 1990
"An 8-bit Josephson Digital Signal Processor," Kotani, Inoue, Imamura, Hasuo, 1990
"The Design for a Josephson Micro-Pipelined Processor," Harada, Hioe, Takagi, Kawabe, 1994
If full 64-bit supercomputers were to be built on this technological base any time soon the
design would necessarily consist of multiple devices for each CPU. Extremely dense packaging
using 3 dimensional approaches would be mandatory to reduce the latency of intraprocessor
communication paths. -— —
Die size for 4-bit processor above:
2.2 mm X 2.5 mm
Free space propagation delay: Propagation in medium with
Gate delay in 4-bit processor:
3.4 ps per millimeter 8.7 ps dielectric constant of 3:
5.87 ps per millimeter
Propagation delay across chip:
8 ps / millimeter
10. Conclusions
This paper describes several architectural concepts which could be dramatically advanced
by three-dimensional packaging. The most intriguing of these is probably the
exploitation of data flow architectures by building large content-addressable memories
for token matching. Fine-grained MIMD systems, while not as general purpose as one
would like given the relative immaturity of today's software, will probably be the most
cost effective way to achieve tera-ops of performance for those algorithms which do fit.
Cache has been the main ingredient to drive microprocessor scalar performance up to,
and even beyond, traditional supercomputers. But the next performance leap will surely
have to come from multiprocessor systems of these devices. Cache coherency that scales
well will certainly be on every designers "to do" list. Very low latency must be offered
by many tightly coupled devices. Area array pads and three-dimensional interconnect
would greatly increase DRAM bandwidth. Any very high speed device technology which
does not support VLSI and ULSI densities will only deliver system-wide high
performance via a very dense interconnect scheme.
11. References
Dally, W.J. and Seitz, C.L. (1987) Deadlock-free message routing in multiprocessor
interconnection networks, IEEE Trans. Computers 36, 547-553.
Kessler, R.E. and Schwarzmeier, J.L. (1993) CRAY T3D: a new dimension for Cray
Research, in Proc. COMPCON 1993, pp. 176-182.
Hill, M.D. and Smith, A. (1989) Evaluating associativity in CPU caches,
IEEE Trans. Computers 38, 1612-1630.
Hill, M.D. (1987) Aspects of cache memory and instruction buffer performance,
Ph.D. Thesis, Univ. of California at Berkeley, Tech. Rep UCB/CSD 87/381, p. 489.
Temam, O., Fricker, C, and Jalby, W. (1993) Impact of cache interferences on usual
numerical dense loop nests, Proc. IEEE 81, 1103-1115.
123
Chaiken et al. (1991) LimitLESS directories: a scalable cache coherence scheme,
in Proc. 4th Intern. Conf. on Architectural Support for Programming Languages
and Operating Systems, pp. 224-234.
Lenoski et al. (1991) Overview and status of the Stanford DASH multiprocessor, in
Proc. 5th Annual ACM Symp. on Principles of Distributed Computing, pp. 229-239.
Tanoi, S., Tanaka, Y., Tanabe, T., Kita, A., Inada, T., Hamazaki, R., Ohtsuki, Y.,
and Uesugi, M. (1994) A 32 bank 256-Mb DRAM with cache and TAG,
IEEEJ. Solid-State Circuits 29, 1330-1335.
Dennis, J. (1979) The varieties of data flow computers, in Proc. 1st Intern. Conf.
on Distributed Computing Systems, pp. 430-439.
Lee, B. and Hurson, A. R. (1994) Data flow architectures and multithreading,
Computer 21, August 1994, 27-39.
Gray, J., Naylor, A., Abnous, A., and Bagherzadeh, N. (1993) VIPER: a VLIW integer
microprocessor, IEEEJ. Solid-State Circuits 28, 1377-1382.
Kotani, S., Inoue, A., Imamura, T., and Hasuo, S. (1990) A subnanosecond clock
Josephson 4-bit processor, IEEEJ. Solid-State Circuits 25, 1518-1525.
Harada, Y., Hioe, W., Takagi, K., and Kawabe, U. (1994) The design for a Josephson
micro-pipelined processor, IEEE Trans. Appl. Superconductivity 4, 97-106.
PROCESSOR PERFORMANCE SCALING
G. A. SAI-HALASZ
IBM, T. J. Watson Research Center
Yorktown Heights, N.Y. 10598
Technology dependent trends are projected for high performance processors. There
are opposite demands placed on the system's area stemming from a need to reduce the
proportion of interconnection capacitance and to send signals across the processor.
Delays resulting from wiring capacitance decrease if processor area increases, while
signal travel considerations favor reducing area. This trade-off for bipolar processors
is governed by power density, while for CMOS the processor size primarily is
determined by wiring considerations. Judicious planning of interconnections to avoid
a so called "RC crises" is necessary to achieve the potential inherent in a technology.
The performance limits of bipolar and room temperature CMOS uni-processors are
very similar. The highest performance existing technology is liquid nitrogen
temperature CMOS. It is not obvious how alternate technologies will fit into the picture
of future general-purpose high-end systems.
1. Introduction
All high performance processors have some common properties stemming from the
requirements that they have to be simultaneously complex and fast. Performance is
measured in million instruction per second (MIPS). MIPS has two components, cycle
time (CT) and cycles per instruction (CPI): MIPS = l/(CTxCPI). The lower a CPI
desired, the more complex the CPU must be. From a practical point of view it means
utilizing a vast number of circuits. But, complexity has its limits too. If the addition
of more circuits hurts CT more than the improvement achieved in CPI, then it is clearly
a loosing trade-off. Also, designing ever more complicated systems may become too
complex a problem for worthwhile pursuit. Hitherto circuit count in high performance
machines has been increasing steadily, a trend which at the uni-processor level is not
expected to continue.
Predicting the CT of a high end processor is actually becoming simpler with
technological advancement. The simultaneous requirement of complexity and speed
leads to certain general constrains that must be obeyed. High circuit count requires
power and large area, while signal propagation favors short distances. These two
opposites lead to some well defined optimal configurations, which are relatively
insensitive to details.
In the case of complex bipolar mainframes it has been well established how
much hardware is needed to achieve a state of the art CPI. Taking this as guidance,
we have chosen to analyze a processor which in its bipolar incorporation consists of
4x10s, 3-input emitter-coupled-logic (ECL) cells, including the peripheral circuits for
the memory arrays. It is also assumed that this CPU has a total of 2Mb of static
random access memory (SRAM) for cache, various directories, translation tables, etc.,
and approximately 0.5Mb of read only storage (ROS). The SRAM cells, even in the
bipolar machines, are assumed to be 6 device CMOS cells; otherwise, due to the low
density of bipolar memories the CPU would have to do with a much smaller cache.
Finally, it is assumed that such a processor has 800 signal input/output ports
communicating with the outside world. For the CMOS machine, the system has an
identical design, and same CPI to the bipolar one, except that the logic operations are
carried out by circuits containing MOSFETs rather than bipolar transistors.
To model cycle time, one has to decide how many logic stages to fit into one
cycle. Since in general purpose systems the pipeline is usually 3 to 5 deep, we are
taking 12 CSEF stages worth of logic as one cycle, which is consistent with breaking
the logic path into about four segments. The exact number 12 is not a critical factor.
A different number of stages, anywhere between 10 and 15, to first order would only
lead to a uniform up or down shift in all of the presented trend curves in the figures.
A generic critical path has additional components besides circuit delays. Delays
due to long wires on the chip, and package delay have to be taken into account as well.
Accordingly, in the modeling, 11 of the 12 stages drive the average wire lengths of the
critical path, while one circuit drives a wire of chip-edge length. The delay through
the 11 ECL stages, interconnected by average wires, plus the 1 driving a long wire, is
the "on chip" portion of the cycle time. A "package" delay to be added consists of two
terms. One is propagation through the package, which we take as having good quality
127
Long wire
length
To arrive from the nominal cycle time to an actual one, one should add the time due
to various tolerances, clock skews etc.. These however are not addressed here although
are critically important. They depend on considerations which are different than those
the rest of this paper is dealing with, and would deserve a whole separate treatment.
However, with proper attention devoted to tolerances they can be expected to scale
with technology in a similar manner as the nominal cycle times, and in percentage
terms to remain roughly constant.
Before introducing the modeling and presenting results it will be useful to discuss
two particular topics which are of importance in understanding performance issues.
One has to do with the physical size of the processor, the second relates to the
importance of dealing with interconnect resistance.
128
With advances in technology leading to smaller features and more powerful devices,
processors are becoming simultaneously smaller and faster. As a result of this trend it
is commonly assumed that smaller area, or equivalently higher circuit density, by itself
results in less circuit delay. In reality this is not the case. To understand this, one must
look at the role of circuit density alone, and not as an accompanying feature of
technology improvement. It is then not difficult to see that under such conditions
increasing circuit density, in fact, increases delay. Delay stems from the time it takes
the current of the driver circuit to charge the combined capacitance of the receiving
circuits and interconnections. Circuit speed improves with increasing area because for
optimized circuit layouts the current delivered by the driving circuit is proportional to
area, while the interconnection part of the load goes up only as the square root of the
area.
The reason that current scales with area is a different one for ECL and for CMOS
circuits. For ECL circuits the current is proportional with area because the power level
of their operation is determined by the thermal energy (heat) that can be removed from
the chip. If a circuit occupies X^m2, and the cooling capability is Y watts/cm2, then
the average power at which the circuit can run is XxYxKT5 mW. The output current
is proportional to the power level. Since for ECL circuits most of the capacitance is
due to wires and not the devices, loading scales as X1/2, giving a delay proportional
with X~1/2.
Let us now look at the situation with CMOS circuits. One expects that in the
high-end arena reasonable cooling capability will be affordable, in which case CMOS
circuits are not power density limited. The reason that the delivered current is linearly
proportional to the circuit area comes from layout considerations. Suppose we have a
maximum area of X available for laying out a certain circuit, and we end up with an
average device width of x. Let's now assume that the available area for the same circuit
changed to X. What we will find is that to the first order the average device width
has changed to x. The net is that, at given design-rules, area directly translates into
device width. Since delivered current is proportional to device width, circuit current
is proportional to area. The argument for the wire loading is the same as for ECL
circuits. Since CMOS does not suffer from the power density limitations of the bipolar
circuits, it is easy to reach a point where the wire load is less than the loading arising
from the receiving circuits. At this point the intrinsic, unloaded, speed of the CMOS
circuit places a floor under delay improvements. Circuit delay as function of processor
area goes as (-area"1'2 + D), where D stands for the intrinsic delay of the circuit. With
proper power expenditure, in high end processors, one can get to the situation where
D becomes the dominant term. At this point decreasing circuit density leads to no
further delay improvements.
The advantage gained by increasing area is only part of the picture. The other
has to do with signal time-of-flight delays. The reason we'll not increase processor
areas indefinitely, is that we have to be able to communicate across the system. When
one is faced with cycle times of only a few nano-seconds, time-of-flight delays across
129
the CPU eventually render area increases detrimental to performance. The next section
deals with this aspect of the systems.
The delay stemming from wire resistance, commonly referred to as RC delay, to first
order is: RxLw x(Q+0.5xCxL w), and has to be included with other delay components.
Here, R and C are the resistance and capacitance of unit wire length, Lw is the wire
length, and Q is the load at the end of the wire. The part in the RC delay due to the
wire alone does not decrease in spite of scaling to smaller dimensions[2]. The factor
that improves the 0.5xRxCxLw 2 term through shorter Lw is negated by the increase
in R due to wire cross section shrinkage. Wire capacitance in the meantime remains
constant, around 0.2pF/mm for minimum width wires with oxide dielectric[3]. As long
as one is dealing with cycle-times over 3-4ns, the wire RC delays are barely noticeable.
However, as we'll see later, for ultra high performance CPUs the resistance in the
wiring can be important for ECL systems, and critical for CMOS processors. The use
of repeaters to regenerate the signal along the way helps, since it decreases the
dependence of interconnect delay on wire length from square to linear. But repeaters
alone do not solve the problem. Delays would still remain unacceptable, and repeaters
entail additional power consumption and design complexity.
If one looks carefully at roles various interconnects play, a better approach
suggest itself[ 1 ]. High performance processors need two kind of wires. First, there are
the wires that serve the vast majority of interconnects. Let's call them "short" ones.
For CMOS processors they are typically up to l-2mm in length. They are mainly
responsible for making the chip "wirable" by providing sufficient number of
interconnections. Here the RC delay plays no appreciable role. Such "short" wires
should follow the minimum lithography features of the available technology. Second,
there is a need for "long" wires, where density is secondary to delay considerations.
They run between distant parts of the chip, and their characteristic length is that of a
chip-edge. A good scaling gauge for such "long" wires is that the time of signal
propagation on them should be only a small fraction of the cycle-time. From such
considerations immediately follows that the cross section of these wires and insulators,
cannot follow minimum lithography features. This type of interconnects will be
referred to as "fat" wires. Fig. 2 shows in cross section an example of the
interconnection scheme needed by ultra high performance processors. It features a
hierarchy of 3 x-y wiring level pairs. The bottom 2 levels are at the finest pitch of
which the device technology can take advantage of. Here lines and spaces should be
almost at minimum design rules. The next 2 levels' dimensions already pay attention
to the RC problem, and finally the top 2 can serve to run signals to full chip-edge
length, or longer, distances. With this type of wiring, where conductor and dielectric
cross sectional dimensions are scaled together, capacitance per unit length stays
constant for each level, while resistance decreases proportionally with wire cross
section increase. In Fig. 2, RC in the second x-y plane-pair is l/4th, and in the third
l/36th, that of the bottom plane.
130
■
Chip
One consequence of having low RC wires is that one will observe transmission line
characteristics not only on the package, but also on the chips themselves. With an
oxide insulator the minimum delay that a signal can achieve due to the finite velocity
of electromagnetic wave propagation is ~7ps/mm. For example, on a 15mm long wire
the signal flight time cannot be less than 105ps. This is significantly longer than the
switching time of drivers in the considered technologies. When the input of a wire is
driven with a faster signal than the travel time down that line, delays are necessarily
dominated by transmission line characteristics, and finite signal propagation speed must
be taken into account.
The net result is that with the proper kind of wiring one can avoid a so called
"RC crisis". The "fat" wire scheme reduces the problem to coping with time-of-flight
delays, which for CMOS at least is a much less severe restriction on performance than
the RC delay would be.
The modeling is described in detail in referenced]. Here only a few of the more
important point will be sketched.
It might seem that the wiring scheme depicted in Fig. 2 wastes too many wiring
channels in comparison with having all levels at minimum dimensions. The difference
between the number of wiring-channels offered by the two cases is not as large as it
might first appear. The reason is that wiring levels block one-another. Thus, a "fat"
wire on the top provides less wiring capability than one at fine pitch, but if the fine
pitch wire were fully utilized, it would impact more severely the number of available
wiring channels in all of the lower levels. If all the pitches are identical, it is estimated
that a level blocks -12 -15% of the wiring capacity of every layer underneath it. This
means that one cannot make systems indefinitely smaller by adding wiring levels.
There exists a minimum size defined by the interconnections. Indeed, if one uses
131
efficient wiring protocols and is able to make good use of the available tracks, then 6
or 7 wiring levels are about the maximum useful number.
Besides wiring, the simulation deals with the number of circuits, the size of the
memory arrays, the number and sizes of the chips, the interconnection capacitance and
resistance, the signal propagation speed in the package, input-output needs, and the
time penalty for chip crossing. Circuit timings are done with simple, linearized,
equations. The coefficients of these equations capture the essential properties of device
scaling. For CMOS effective FET resistance in digital applications is best
characterized by the device's large-signal-transconductance[5]. Since
large-signal-transconductance has already been experimentally obtained[6,7] for
devices from 0.28pm gates down to 0.07um, one does not need to explicitly deal with
intrinsic device parameters, such as mobility, velocity saturation, and the like.
The capacitive load and wire length in the critical paths are obtained from the
average net length. Resistive and time-of-flight delays, calculated separately, are
combined with the capacitive and intrinsic delay components.
The two main indicators in the degree of advancement in a bipolar ECL processor are
the power density, and the speed of the bipolar transistors at a given current level.
1.6 —I 1—
10OW/cm2
(0
J 1-4 1-
Ü
>
Ü
B
C
I012
z
200
1.0
0.5 1 1.5 2 2.5 3
Chip Area (cm2)
Fig. 3. "Ultimate" bipolar ECL CPU embodying. Power densities and device power
in critical paths as indicated.
132
Fig. 3 and Table 1 show the type of bipolar ECL system which one can regard as an
ultimate goal. Device speed is characterized by 14ps delay at 5mW and a 400mV
signal swing. Lithographic capability is 0.25pm and we assume that the CPU occupies
4 chips. Clearly, such advances can translate into system speed only if power density
is increased as well. The initial sharp delay improvement in Fig. 3 with increasing area
is the result of the fundamental behavior of area vs speed we discussed earlier. The
more interesting parameters of the 200W/cm2, 2cm2 chips ECL processor are given in
Table 2.
Table 1. Characteristic parameters in a possible "ultimate" bipolar uni-processor.
Parameters/Characteristics "Ultimate" bipolar CPU
CPU arrangement 4 chips, 2cm2 each
Circuits/Technology ECL, SiGe transistors
14ps unloaded delay @ 5mW
Signal swing 400mV
Lithography 0.25pm
Power density in logic 200W/cm2
Circuit power in critical paths 5.6mW
Wiring 4 levels: 2 @ 1.6pm pitch
2 @ 6.4pm pitch
Average net length in in critical paths 0.9mm
Average stage delay in critical paths 38ps
On-chip long wires and package delay 500ps
Nominal cycle time Ins
Logic power consumption 1550W
Even at 200W7cm2 the system is not capable of making use of the density that
lithography would allow. Intrinsic device speed has little to do with ultimate system
performance. Indeed, even if in this processor we were to assume zero intrinsic device
delay, nominal cycle time would improve only -15%. Most delays stem from the size
of the CPU, which due to power density limitations, cannot be shrunk sufficiently.
The problem is that in bipolar ECL circuits steady current is flowing between voltage
levels that are related to the silicon bandgap, and thus must stay roughly constant in
spite of miniaturization.
4. CMOS Processors
To break the barrier of power density limitation, a technology is needed where power
consumption is limited to charging/discharging of capacitances during logic operation.
It is also important to be able to decrease power supply voltages along with dimensions.
CMOS is such a technology.
Modeling performance in a CMOS processor is more difficult than in a ECL one.
Complications primarily arise in calculating wirability for the processors. Wiring
statistics[8] involves some empirical factors, which for CMOS processors are not yet
well established. Theoretically, the Rent exponent for random logic it is 2/3, while for
memories it is about 1/2. For high density CMOS, where a whole processor, made
up of functional blocks having relatively little communication with each other, can be
integrated on a single chip, the Rent exponent is not well established. We assumed that
a processor consists of a small number of building blocks within which a the Rent
exponent is high, but the interconnections between the blocks is governed by a smaller
exponent. Many cases were investigated to determine what influence such choices have
on performance. As it turns out, they have practically no impact. However, they
133
(A
C
■^
fl)
Ü
>
0
«
c
£
o
z
0.4
0.3 0.2 0.1 0.05
Gate length (jam)
Fig. 5. CMOS high-end processor scaling path
With this introduction we can now look at the expected CMOS CPU performance
progress with improving technology given in Fig. 5. The nominal cycle is shown as
function of gate length. The gate, instead of the channel, was chosen as measure
because it is a straightforward indicator of the state of lithographic capability at any
given time. For both the RT and LT processors, the overall lithographic ground-rules
are assumed to be more relaxed than the gate length itself. Progressing with a single
lithographic level ahead of the overall capability is technically entirely feasible. The
135
density of device gates is quite small, which means that even electron-beam lithography
can be used if the shortest gates were problematic by other lithographic methods. In
Fig. 5, at the 0.25pm gate length symbols 0.5pm overall process ground-rules were
assumed. The finest wiring pitch was set to to 1.2pm with 0.6pm of lines, spaces, and
interlevel separations. At the 0.15pm gate length point ground-rules are 0.25pm, and
minimum wire pitch 0.6pm. Such values were chosen as the most probable ones based
on how technology is likely to progress.
The RT improvement between the 0.25 and 0.15pm points is a factor of -1.8,
with the 0.15pm gate length cycle time approaching 1.5ns. The improvement results
mainly from faster devices as gate length shortens, and as gate oxide goes from 7nm
to 5nm. Junction capacitances are also significantly reduced with ground-rules
shrinking by a factor of two. With finer wiring the chip size can be reduced by almost
by a factor four, and time-of-flight decreases.
To give an indication of the weight with which various parameters influence
performance we take a detailed look at an intermediate point: 0.18pm gate length and
0.35pm ground-rules with 2.0V power supply. This processor has 1.8ns nominal cycle
at 55W of power consumption, on a 2.5cm2 size chip. Table 2 shows cycle time
sensitivity to some parameters.
We are looking at a processor optimized for performance. All the devices are
of maximum width limited only by area considerations. In the critical path circuits
loading comes 65% from devices and 35% from wires. The circuits are only 25%
slower than their intrinsic unloaded speed. For such a system gate length is the single
most important parameter: a 10% change results in 8% cycle time improvement. Since
wire loading is minimized already through the use of maximum width devices, 10%
improvement in the dielectric constant would bring only 2.5% return in cycle time.
However, the dielectric constant can be leveraged to save power. If one uses the same
10% dielectric constant reduction to remain at the original speed, but with narrower
devices, one obtains a 15% power reduction. Finally, we come to wire resistance.
Performance insensitivity to this parameter results from the use of the "fat" wire
scheme. If we did not have the low RC interconnects, but only 6 layers at the
minimum pitch fitting this device technology, about 0.85pm, we'd have suffered a 60%
performance loss, even if we had been allowed to use very wide wires when needed.
Also, adding to complexity, the long wires would be in need of 4 to 6 repeaters,
without which cycle time would deteriorate a factor of about 2.5.
Returning to the discussion of Fig. 5, below 0.15pm gate length we ceased improving
ground-rules, keeping them at 0.25pm. There are several reasons for this. First, true
0.25pm lithography is already stretching the limits of technology, and to go below
0.25pm would be difficult indeed. Second, there is not much to gain by shrinking
ground-rules further. Third, progressing below 0.25pm can actually be detrimental to
performance. The reason is contact resistance. Even a contact resistance value close
to theoretical limits, like 2xl0"7£2/cm2, which is not likely to be ever attained in
millions of contact holes in the CPU, is a significant factor at 0.25pm. Shrinking
contacts further counteracts other advantages derived from a smaller system.
136
On Fig. 5, the last RT gate-length point is 0.15pm. The dashed line continuing
to 0.12pm reflects the view that somewhere around 0.15pm is the RT gate limit in
performance oriented CPUs. Shorter gate-length devices had been made and operated
quite successfully at room temperature. But it is one thing to fabricate a few thousand
devices, and quite another to have a viable design for a system with millions of
transistors, where tolerances play a pivotal role. The main problem with RT operation
is the inability to turn off devices sufficiently. For this reason a relatively high
threshold of -0.5V is the lowest viable at RT. The high threshold then requires a
power supply around 2V. To live with such a high power supply one has to introduce
corrective measures into the device design which are detrimental to performance. And,
performance would have been the reason to shorten gates to begin with. In summary,
in spite of a recent surge in optimism regarding a 0.1pm RT technology, we are
doubtful that in the highest performance CPUs we will see nominal gate lengths much
below 0.15pm. On the optimistic side, let's not forget that such a 0.15pm gate-length
processor would still be many times as powerful as todays mainframes. With some
help from custom circuit design it might reach, or even better, a Ins cycle time, which
would at least equal what an "ultimate" ECL bipolar processor could do.
Finally, lets turn our attention to the LN2 temperature situation. Looking at Fig.
5, LT obviously offers the highest performance of all systems. The performance
advantages of FET LN2 operation have been recognized and advocated for a long
time[9].. However, it appears that until performance improvements can be made at RT,
LN2 operation will remain a matter of discussion only. The aim in presenting Fig. 5
is to contend that shortly LN2 temperature CMOS will have to be taken seriously
because it is the only avenue open toward higher performance processors. As apparent
from Fig. 5, LT offers 1.6-1.7 times the performance of a RT processor in the same
technology generation. This is mainly due to better device properties, and somewhat
to the lower, -1/5, wire resistance. The main point, however, is that at LT gate lengths
can be shortened at least down to 0.07pm[7] with spectacular performance gains. The
fundamental reason for scalability of FETs at LN2 is that devices can be turned off
much more readily than at RT. This fact allows for a whole different low threshold,
low voltage design space from which RT operation is excluded. Comparing the room
and low temperature performances at their probable limits of 0.15 and 0.07pm
gate-lengths, we see that LN2 temperature offers an almost three fold advantage over
what is achievable at room temperature. It is worth contrasting the parameters of these
two systems, and comparing them with the Ins cycle-time ECL bipolar processor.
Table 3 summarizes the important parameters of the 0.55 Rent exponent scenario.
Table 3. Selected parameters of possible "ultimate" CMOS processors
at room and LN2 temperatures
Parameters/Characteristics Room Temp. CMOS CPU LN2 Temp. CMOS CPU
CPU with optimistic Single, 0.8cm2 area chip Single, 0.8cm2area chip
wiring assumptions
Overall lithography design rules 0.25pm 0.25pm
Gate lithography 0.15pm 0.07pm
Gate oxide 5nm 2.8nm
Power supply 1.8V 0.8V
Wiring 6 levels; 2 @ 0.6pm pitch 6 levels; 2 @ 0.6pm pitch
("fat" scheme) 2 @ 1.2pm pitch 2 @ 1.2pm pitch
2 @ 2.4pm pitch 2 @ 2.4pm pitch
Average net length in critical paths 80pm 80pm
Average stage delay in critical paths 40ps lOps
Cycle time portion caused by long wires 300ps 220ps
Nominal cycle time 1.5ns 520ps
Logic power consumption 18W 10W
137
In this technology at LN2 a logic stage takes only ~10ps; hence the long wire delay
plays proportionally a more important role than in slower technologies. Unfortunately,
contact resistivity plays a larger role as well. If it could be eliminated it would lead
to ~90ps, more than 15% cycle-time improvement with the 0.25um design-rules.
Furthermore, without contact resistivity problems it would be worthwhile to reduce
design-rules to, let say 0.15pm, and one could have a 370ps cycle-time processor on
a 0.4cm2 chip at 10W.
5. Discussion
It has been shown that performance directions at the high end are quite predictable.
The time of CMOS in the performance arena has arrived. However, performance
potentials are not limitless. A combination of obstacles, mainly due to device behavior,
are identifiable.
At the chip size where a given CMOS processor becomes just wirable total wire
capacitance scales with design-rules, while total device capacitance scales with
(design-rules)2xCox. This means that eventually CMOS too, could find itself
overwhelmed by wires. However, this would occur only at dimensions an order of
magnitude below the probable device scaling limits. Accordingly, well designed, high
performance CMOS processors should be immune to wiring imposed limits. Since
wire capacitance per unit length does not scale with lithography, any technology which
tries to replace CMOS by shrinking dimensions beyond those of CMOS, will have to
face the wire loading problems, and provide the necessary current drive capability to
deal with it. We are not aware of any such technology on the horizon.
Contact resistivity between a semiconductor and metal, as discussed earlier,
becomes a performance crippling effect at deeply submicron dimensions. Again, any
semiconductor technology would have to surmount this obstacle. Along these lines, a
deeply scaled Schottky source/drain CMOS might be the path out of contact resistance
difficulties.
One has to say a few words about the communication needs of high end
processors, either in stand-alone, or in multi-processing configurations. To make full
use of future processor performance a communication bandwidth of several hundred
Gbit/s should be provided, together with memory capacity measured in Tbits. The
communication angle is one where optical techniques based on compound
semiconductors could find probable applications.
The considerations above relate to general purpose processors. If one is satisfied
to serve only one specific task, for instance, a machine dedicated to matrix inversion,
or for data searches, performance for that particular application can probably greatly
exceed the ones arrived at in this article. Hardware being ever cheaper, one can
envision the end of general purpose machines in the highest end domain. The obstacle
is the time spent and the difficulty of system and circuit design. It is worth to consider
a research effort into design systems, which could result in fast delivery of a systems
dedicated to a specific task. This may ultimately be the most cost effective path toward
performance beyond the present projections.
In summary, the performance potentials at both RT and LT CMOS, coupled with
the relatively low cost of FET technology, is quite exiting. The real challenge will be
for society to absorb what such technologies can deliver.
6. References
Herb Goronkin
Motorola
2100 East Elliot Road
EL508
Tempe, AZ 85284
1. Scaling of VLSICs
So let's start with some predictions that fall into the category of being
nearly safe. In figure 1, the transconductance of heterostructure FETs and
silicon MOSFETs is plotted against the gate length. All of the data are
taken from the literature. Even if we ignore the underlying physics of the
trends of the data, we can predict that when the gate length is less than
about 70nm, silicon FETs will have larger transconductance than III-V
HFETs. The underlying reason for the superior scaling of silicon is the
dependence of transconductance on gate insulator thickness and the
dependence of tunnel current on both the energy gap and thickness. The
tunnel current is proportional to
139
S. Luryi et al. (eds.), Future Trends in Microelectronics, 139-149.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
140
exp(-2
V 2m*AE
h2 Tox) (1)
where AE is the majority carrier band offset and Tox is the insulator
thickness. The transconductance, gm, is approximated by
evZ
gm= Tox (2)
where v is the channel velocity and Z is the gate width. Figure 2 shows
the scaling trend of gate length and gate oxide thickness based on
extrapolation of introduction dates of DRAM families. In order to
maintain approximately the same magnitude of transconductance as the
gate width shrinks and in order to avoid short channel leakage as the gate
length scales, the gate oxide thickness must scale with the gate length. In
silicon, for gate oxide thickness below about 3nm and supply voltage of
about 1.0 volt, the tunneling current will be several orders of magnitude
larger than the channel current and the gate cannot control the channel
charge. In III-V HFETs the insulator thickness must be larger to
compensate for the much smaller AE which is on the order of 0.3 -1.0 volt
compared to 3.5 volts at the silicon-silicon dioxide interface.
1
E \
£ \
35 ^ ■
E v %
Ni. **»
0)
\ ^v
>B.
**> *■*
^
O 1000
ra L ^W ^
o VlnP/lnGaA^N^^^ ^ ^inAs
3
T> \ ^^^**"***^H^ >W ^S^'Nw
C ^^^ ^^**H**H*»fc^ ^**N**W ^^^^
o
o ^^^^^ ^^**%^^\»
in Si NMOS^S. GaAs/lnGafi^O\^
c .
is
!■_
H
.
0.1 1
Gate Length (pm)
Figure 1. Trends of experimental
transconductance as a function of gate length for
FETs of various materials.
141
1 0
16M
64M
1 0°
>56M
CO _1G
1 0'
Number of Electrons 16G
u 1 64Gb
E «> io 256Gb
3ED LTb
Gate Length
Oxide , Thickpess
1 0
1988 1992 1996 2000 2004 2008 2012 2016 2020
Year
Figure 2. Extrapolated historical trends of DRAM scaling.
Using Figure 2, we can also predict that the number of electrons in the
DRAM channel will decrease as the density of transistors increases. This
arises from the need to keep the chip from melting during operation and to
maintain a reasonable chip cost.. The projected number of electrons in the
64Gb DRAM channel is about 10. We might expect somewhat unusual I-
V characteristics as one or two electrons, which represents a large fraction
of the total number of electrons, enter or leave the channel. Let's examine
this possibility in a little more detail.
Figure 3 illustrates the concept of Coulomb blockade. When the device and
associated capacitance are sufficiently small, an electron entering the
142
fc-c (3)
where E is the energy and C is the device input capacitance. The value of
C determines whether the I-V characteristics will be smooth as is in
classical transistors or stepwise. Accordingly, if we consider the normal
operating range of 0-80C, E must be greater than 4kT giving the result that
C must be less than 0.6aF. At this temperature a larger value of
capacitance will cause the steps to become smeared and continuous. For
practical device operation a more stringent condition of 40kT should be
used to achieve device stability.
_JQL.
q/2C^
Coulomb *"T"
Before Filling /2C
Blockade <l 1
Figure 3. The electron transferred into the quantum well raises the energy
level by an amount q2/2C and a voltage equal to q/2C must be applied to
pull the eigenstate back into alignment with the electron gas in the
emitter.
Using the scaling trend, let's examine whether the 64Gb DRAM is likely
to exhibit normal or step-like I-V characteristics at normal operating
temperatures. Figure 4 shows the gate capacitance, number of electrons
and corresponding DRAM density as a function of gate length. From the
capacitance requirement for 300K operation the arrow points to the
maximum value of capacitance for stepwise current increase. Continuing
to use the 64Gb DRAM as an example, although the number of electrons
scales to about 10, the capacitance scales to about 70aF. In order for clear
single electron behavior to be observable, the temperature needs to be
reduced to 3.3K. Therefore, we can predict that if traditional scaling
continues, the 64Gb DRAM will have smooth I-V characteristics.
143
104 I I I I I I I I I 104
LL W 10 Ö
Soio3 33
© -K
ü u 102 <£
jS HI 10 Ö
101 §
CO O
c/>
ü .Q -. 10
E
Ä Q
CO = 10° T 10
1 cr
CD 2
Single Electron 300K Operation
.-1 i i i i ' ' ' ■ * i i ■ « » ■ ■»
10 10"
1 o 100 1000
Gate Length (nm)
Figure 4. Scaling of gate capacitance and number of electrons with
DRAM density. The arrow points to the maximum value of capacitance
for observation of sharp stepwise I-V characteristics.
AN 1
N :
N (4)
I I
ID ±10% ID ±10%
AVL
Density 1Tb
Voltage 0.25V
Power dissipation 10W
Access time 10ns
Max operating temperature 80C
Chip size 2.5x2.5 cm
If we suppose that scaling can take us to the CSI regime, the above
discussion leads to the conclusion that in about 2010 we will have to have
a nanometer-scale fabrication technology in manufacturing labs in order to
meet the introduction date of 2016. Since the electron density scales to 2
electrons per cell, conventional transistors will have to be replaced by a
more deterministic structure. In addition, with scale lengths on the order
of ones to tens of atoms, there is presently no conceivable lithography that
can perform pattern transfer to the accuracy of atomic dimensions.
Channel Atoi
Figure 6. Transfer of charge in gate circuit retards or allows charge
transfer in channel.
1.
q/2C IV
The gate can contain one or two electrons. If the gate contains one electron
in the position nearest the channel, a channel electron can repel the gate
electron. The second electron holds the bias electron in place. The bias
electron raises the second atom state and this retards the flow of charge
into that stage. We can derive a transconductance for this structure
(5)
This section is for handwaving. It is here that the problems of using the
smallest transistor begin to emerge and the solutions to those problems
have not been adequately addressed. So I will make some predictions and
hope they will serve as a starting point for further discussion that may
actually lead to useful approaches.
6. References
Likharev, K.K. (1991) Single Electronics: Correlated transfer of single electrons in
ultrasmall junctions, arrays and systems, Granular Electronics. D.K. Ferry, J.R. Barker,
C. Jacobini, Eds. New York: Plenum, 371-391
2
see for example, Tougaw, D. P., Lent, C.S. (1994) Logical devices implemented using
quantum cellular automata, J. Appl. Phys 75 (3), 1818-1825
Challenges and Trends for the Application of Quantum-Based Devices
1. Introduction
A major challenge in the application of quantum-based devices arises from the molecular
feature-size considerations implicit in nanoelectronic and mesoscopic technology;
indeed, many interesting questions arise concerning fluctuations, tolerances, robustness,
and other statistical considerations which might conceivably wash-out many of the
seemingly fragile characteristics of nanodevices. There are many illustrative examples in
which inherent statistical variation in composition and device dimensions produce
substantial deviation from the desired nanostructure electrical response. These examples
include: minimum metal-oxide-semiconductor transistor size as determined by a
combination of gate oxide breakdown, drain-source punch-through and substrate doping
fluctuations [1]; minimum planar bipolar transistor size as determined by a combination
of collector junction breakdown, base punch-through and base doping fluctuations [2];
effects of structural and alloy disordering on the electronic states in quantum wires [3];
151
5. Luryi et al. (eds.), Future Trends in Microelectronics, 151-157.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
152
and the effect of fabrication-related dimensional variations on carrier scattering rates in
quantum wires [4,5].
Clearly, a challenge for nanoelectronic and mesoscopic device communities is to
explore concepts and designs that optimize robustness and suppress device fragility. In
meeting this challenge it will be essential to circumvent statistical phenomena which
typically exacerbate the scaling of conventional device technologies to reduced feature
sizes; as well, it will also be essential to exploit phenomena specific to the nanoscale
regime which are known to suppress noise and related fluctuations in "small" and
mesoscopic devices. On a more positive note, a related challenge for nanoelectronic and
mesoscopic device communities is to explore concepts and designs that optimize
robustness and suppress device fragility.
Conceptual options for achieving robustness are an essential ingredient for the
successful exploitation of quantum-based information processing devices and systems.
Indeed, several phenomena which play critical roles in nanoelectronic and mesoscopic
devices provide important opportunities to realize robustness in such nanodimensional
devices.
Promising avenues for achieving robustness include the application of Coulomb
blockade effects, design through quantum control theory, and emulation of biological
and chemical systems where phenomena such as neuron networking and self-
organization finesse disordering processes. For example, potential applications of
quantum control theory to "small" and mesoscopic electronic devices are motivated
based on past uses of robust optimal control theory for the selective excitation of
quantum mechanical vibrational states of molecules [6-10]. Additionally, a number of
important, recent developments in the field of Coulomb blockade are highly
encouraging; these developments include: the observation of Coulomb blockade effects
at temperatures which are an appreciable fraction of room temperature [11]; theoretical
prescriptions for enhancing the reliability of single-electron switches operating on the
basis of Coulomb effects [12]; and recent progress in understanding how mesoscopic
Coulomb blockade effects may be used to greatly suppress noise in electron emission
processes in p-i-n semiconductor junctions as well as in p-n microjunctions [13]. The
observation of Coulomb blockage effects at 100 degrees Kelvin [11] have been extended
recently by the principal authors of [11] to 110 degrees Kelvin for the case of holes and
170 degrees Kelvin for electrons. As a means of expanding on the progress made in
some of these areas, this paper places particular emphasis on two specialized topics: the
simulation of the capacitance of quantum dots [14]; and the tailoring of deformation
potential and piezoelectric scattering in mesoscopic devices in order to maintain de
Broglie wave coherence [15].
The impressive trends in both the fabrication and design of quantum-based devices
motivate and lead to the need for circuit design tools for integrated circuits with
153
quantum-based component devices. In the recent past there has been substantial
progress in the development of design tools for ultrafast and compact circuits using
heterojunction bipolar transistors and negative differential resistance (NDR) devices such
as resonant hot-electron transistors, resonant tunneling transistors, and resonant
tunneling diodes; examples of these works are summarized in [16]. Indeed, these efforts
are realized, in part, based on a new circuit simulator, NDR-SPICE, which has been
developed by extending the Berkeley SPICE simulator to the domain of NDR circuits.
To date such tools have been used to design a wide variety of circuits including:
multiple-valued multiplexers and demultiplexers; totally parallel multiple-valued logic
adders; four-valued up/down counters; analog-to-digital converters; and a 32-bit parallel
correlator.
Recent successes in defining quantum logic gates that are potentially suited as the
building blocks of quantum logic networks [17-19] portend new vistas and enormous
payoffs through the realization of quantum computers. This possibility has not been
the focus of extensive attention within the microelectronics and nanoelectronics
communities but these communities have created fabrication and design technologies
that may provide the basis for realizing quantum computers.
4. Quantum Capacitance
0.8
0.6 i-D-a
Q
QDO* P°- ü
o-cr
7 12 17 22
Number of electrons
Figure 1. Capacitance versus electron number for a two-dimensional GaAs quantum dot with dimensions as
defined in the text.
plotted as a function of the electron number of atoms with nuclear charge number Z.
These non-classical effects in the quantum capacitance should be important in a variety
of applications of quantum-effect devices including quantum-dot memory cells for
compact high-density data-storage devices.
1000 T
Figure 2. Capacitive energy versus electron number for atoms with nuclear charge Z.
155
6. Conclusion
The challenges before the nanoelectronic and mesoscopic device communities are — on
the one hand — to explore concepts and designs that optimize robustness and — on the
other hand — to exploit quantum effects that open the way to unique means of
information processing that have no counterparts in the classical domain. Through this
dual approach the nanofabrication technology revolution can be exploited fully by the
156
architects of future generations of quantum-based information processing devices and
systems. As predicted by the famed physicist Richard Feynman in his 1959 talk
entitled "There's Plenty of Room at the Bottom," we shall realize the "bottomless"
possibilities of "manipulating and controlling things on a small scale."
7. References
3. Singh, Jasprit (1991) Effect of structural disorder on electronic states in GaAs/AlGaAs quantum
wires, Appl. Phys. Lett. 59, 3142-3144.
4. Mickevicius, R., Mitin, V. V., Kim, K. W., Stroscio, Michael A., and Iafrate, Gerald J. (1992)
Electron intersubband scattering by confined and localized phonons in real quantum wires, J. Phys.
Condens. Matter 4, 4959-4970.
5. Mickevicius, R., Mitin, V. V., Kim, K. W., and Stroscio, Michael A. (1992) Electron high field
transport in multisubband quantum wire structures, Semicond. Sei. Technol. 7, B299-B301.
6 Warren, Warren S., Rabitz, Herschel, and Dahleh, Mohammed (1993) Coherent control of quantum
dynamics: The dream is alive, Science 259, 1581-1589.
7. Shi, Shenghua, and Rabitz, Herschel (1992) Optimal control of selectivity of unimolecular reactions
via an excited electronic state with designed lasers, J. Chem. Phys. 97, 276-287.
8. Beumee, Johan G. B., and Rabitz, Herschel (1992) Robust optimal control for selective vibrational
excitation in molecules: A worst case analysis, J. Chem. Phys. 97 1353-1364.
9. Rabitz, Herschel (1992) Optimal control of molecular motion, in A. D. Bandrauk and S. C. Wallace
(eds.) Coherence Phenomena in Atoms and Molecules in Laser Fields, Plenum Press, New York, pp.
315-331.
10. Dahleh, M, Pierce, A. P., and Rabitz, H. (1992) Design challenges for control of molecular dynamics,
IEEE Control Systems 12, 93-94.
11. Leobandung, Effendi, Guo, Lingjie, Wang, Yun, and Chou, Stephen Y. (1995) Observation of
quantum effects and Coulomb blockage in silicon quantum dot transistors at temperatures over 100
Kelvin, J. Appl. Phys., in press.
157
12. Imamoglu, A. and Yamamoto, Y. (1993) Noise suppression in semiconductor p-i-n junctions:
Transition from macroscopic squeezing to mesoscopic Coulomb blockade of electron emission
processes, Phys. Rev. Lett. 70, 3327-3330.
13. Imamoglu, A., Yamamoto, Y., and Solomon, P. (1992) Single-electron thermionic-emission oscillations
in p-n microjunctions, Phys. Rev. B 46, 9555-9563.
14. Macucci, M., Hess, Karl, and lafrate, G. J. (1995) Simulation of electronic properties and
capacitance of quantum dots, J. Appl. Phys. 77, 3267-3276.
15. Stroscio, Michael A. and Kim, K. W. (1994) Piezoelectric scattering of carriers from confined
acoustic modes in cylindrical quantum wires, Phys. Rev. B 48, 1936-1941.
16. Mohan, S., Mazumder, P., Haddad, G. 1., Mains, R. and Sun, S. (1991) Ultrafast pipelined adders using
resonant tunneling transistors, IEE Electronics Letters 27, 830-831; Mohan, S., Mazumder, P., and
Haddad, G.I. (1991) Subnanosecond 32-bit multiplier using negative differential resistance devices,
IEE Electronics Letters 27, 1921-1931; Mazumder, P. (1994) Picosecond pipelined adder using 3-
terminal devices, IEE Proceedings E, Computers and Digital Technics 141, 104-110.
17. Barenco, Adriano, Deutsch, David, and Ekert, Artur (1995) Conditional quantum dynamics and logic
gates, Phys. Rev. Lett. 74, 4083-4086.
18. Sleator, Tycho and Weinfurter, Harald (1995) Realizable universal quantum logic gates, Phys. Rev.
Lett. 74, 4087-4090.
19. Cirac, J. I. and Zoller, P. (1995) Quantum computations with cold trapped ions, Phys. Rev. Lett. 74,
4091-4094.
20. Stroscio, Michael A., Kim, K. W., Yu, SeGi, and Ballato, Arthur (1994) Quantized acoustic phonon
modes in quantum wires and quantum dots, J. Appl. Phys. 76, 4670-4675.
WIRE AND DOT RELATED DEVICES
Abstract
1. Introduction
n
•2- 2££Q/W
-n2D(j + a)
w
7 = 1,3,5 (1)
with the effective two dimensional carrier density in the wires n2D>the effective mass of
the carriers m* and the electrical wire width w and a, which denotes a correction for the
162
phase shift occurring due to the reflection of the electrons at the wire walls. In the FIR
experiments we clearly identify the plasmon shifted CR with the first odd mode j=l.
In a density modulated system, one expects to find two resonances [7], the
unperturbated CR and an extended plasmon, which is a better description, when coupling
between the electrons in the different parallel wires becomes important. Then C0p is
dominated by the grating period a [8]:
2K
(o„ =■ n (2)
2D-
lee.0*
By increasing the illumination time, the potential modulation disappears at all, and one
finds the pure two dimensional behaviour of the system. By subsequently etching the
sample, the resonance position shifts to higher energies, and is not affected anymore by
above bandgap illumination. Due to
surface charge depletion effects, the
effective electron concentration decreases
inversion channel rapidly and therefore the amplitude of the
CR
resonance decreases also.
In Fig.2 a FIR
photoconductivity spectrum is shown
taken at a laser wavelength of 163|J.m.
Comparing the transmission data with
the data obtained from the
photoconductivity measurements
performed on the same sample, one
clearly sees, that the resonances appear at
the position of the confined plasmon.
sample A The plasmon energy is easily determined
35s etched from a plot like fig.3. by extrapolating
1ms illuminated the point of intersection of the line of
the plasmon energies with the y-axis.
3.0 3.5 4.0 4.5 5.0
Magnetic Field (Tela) In addition, there is a small, but very
sharp resonance (FWHM = 0.2 cm"*' an
Fig. 2 Typical spectra obtained with the FIR-laser in effective mass of 0.067 mo and an
photoconductivity and transmission measurements of
sample A. effective 2D electron density of 4-1010
cm"2) near the position of the unperturbed 2D CR. It can also be seen in non-structured
samples and its position is not affected by the etching process of the wires. So we expect
the origin of this resonance is due to an inversion channel deep in the sample, as well
known from similar grown samples [9].
163
The spectrum taken with the FIR laser is in good agreement with the bolometric model
of the PC [10], [11], where the PC grows both with the absolute resistance of the
system and its temperature dependence. The evaluation of the resonance energies for PC
and transmission measurements of the sample A with a=630 nm and an etching time of
35s are given in a double squared plot (fig.3.) With good agreement between the different
methods, the plasmon resonance follows the quadratic dispersion relation
1 ? o (3)
(o =(Op+ a)*
which is obtained for the collective plasmon excitation of the electrons in an magnetic
field coupling quadratically to the cyclotron resonance energy coc. Assuming an harmonic
oscillator model for the confining potential
yext=\m*C0lx2 (4)
the same relation for the dispersion of the ID resonance energy as eq.(3) can be obtained.
For the non illuminated sample A (35s etched) we derive a subband spacing ©0=3 meV
and an electrical width w = 380 nm (dark) and w = 510 nm (illuminated), calculated from
eq. 1 with e = 12.9. Differences in the resonance position between transmission and PC
values are attributed to changes
12000 M | I I I I | I I I I | I I II | in the population of the wires,
which may have different causes:
10000 sample a
a=630nm heating of the electron system by
8000 35s etched the high FIR intensity as well as
T=4.2K heating and charging of the wires
6000
— non structured by the current.
CD
a> It is shown, that via above band
c 4000
ID o photoconductivity gap illumination the electronic
2000 properties of the shallow etched
• transmission wire structured sample could be
i.... i. ... i .... i .... i .... i ...
changed from quasi ID to a
20 30 40 50 60 70
density modulated 2D system and
Magnetic Field2 (Tesla2)
finally to a pure 2D system. The
Fig.3. Squared energy of the evaluated resonance position PC is strongly correlated to the
(transmission and photoconductivity results) versus squared position of the localized plasmon
magnetic field (sample AThe dotted line corresponds to the
unperturbed 2D CR with a mass of m*= 0.07mo in FIR transmission. Recent PC
measurements [11] were
successfully performed on a tiny sample of 80 x 200 \im2 with 130 wires. The big
advantage of the PC measurements is, that there is no need for a large structured area as
in transmission experiments. One disadvantage of this quantum wire array (some
164
hundreds) is till now the very bad detectivity, but combining holographic and electron
beam lithography, it is possible to prepare photoconductive detectors consisting of very
few wires with much more improved properties.
The samples for 0-2D tunneling consist of a nominally undoped GaAs layer
(NA <lxl015cm-3) grown on a semi-insulating substrate, followed by an AlxGai_
xAs barrier of a total thickness of 200 A [50 A spacer, 50 A doped
(A^D =3xl018cm"3)> 100 A spacer; x=0.36], and a n-doped GaAs layer [d= 800 Ä,
ND = 1.2xl015crrT3]. An additional GaAs cap layer was highly n-doped [d= 150 A,
JVD = 6.4 X 1018 cm"3]. This structure provides two 2D systems separated by a barrier
of only 200 A. From Shubnikov-de Haas measurements it was deduced that in both
166
2DEG systems only one subband is occupied having electron concentrations of
nim = 6.0 X1011 cm"2 and <cc = 5.5 x 10" cm"2, respectively.
AuGe - contact
etched area
2DEG
AlxGai-xAs
Fig.6 : A schematic view of the sample is shown on the left hand side. The corresponding conduction band
profile for the etched and non etched areas of the samples is shown on the right.
The sample geometry is shown in Fig.6(a). First, bar shaped mesas were etched and
Ohmic AuGe contacts were aligned to both channels. Then, a holographic photoresist dot
patterns with a period of a=350nm was fabricated on the mesas, which were etched wet
chemically into the GaAs cap layers. AuGe was evaporated on the total area of the
quantum dot system. Finally, the GaAs layers around the top tunneling contact were
removed selectively, yielding independent contacts to both the multiple quantum dot
(MQD) system and the 2DEG system. By applying a voltage Vt>, the 0D states are
shifted energetically by AE = eAVb with respect to 2DEG [17] The bandstructure of the
sampleis shown in Fig.6(b) for both the etched and the non-etched regions (upper and
lower part, respectively). All measurements were made using a 4-terminal conductance
bridge [18]
The experimental results are plotted in Fig.7. The lower part (a) of this figure
shows the dl/dVb characteristics of the nano-structured (0D-2D) sample, where the
temperature is varied between T=1.7K [curve (1)] and T=40K [curve (12)]. For reference
reasons, the dl/dVb characteristics of a not-nanostructured sample are plotted in the upper
part (b) of Fig. 7 for the two temperature values T=1.7K [curve (1)] and T=40K [curve
(2)]. The comparison of the two characteristics [a(l)] and [b(l)] shows that the nano-
fabrication process leads to a multitude of new resonances which exist within the whole
167
voltage range considered. All resonance peaks of the (0D-2D) dI/dVD characteristic show
a strong dependence on the temperature. Above T=4.2K [curve a(2)], only about half of
the resonances can still be resolved. A further increase of the temperature results in a
monotone broadening of all resonance structures, accompanied by a monotone decrease of
peak amplitudes.
The tunneling probability for transitions between the subbands of the 2DEG system
and the fully quantized states of the MQD system are calculated in analogy to the the
selection rules for 1D-2D tunneling processes [19] Details are published elsewhere [20]
To compare the calculated results with the experiment, the overlap integral between the
wave functions on the 2D and 0D side were calculated as a function of bias voltage. The
overlap integral is directly proportional to the tunneling current. A cosine shaped
potential best describes the experimental situation in the dots and the main peaks in the
wave tunneling probability are always expected when w=ijr. Ik where i is an integer, k is
the wave vector of the tunneling electron and w is the width of the potential at the
coresponding energy. Note that the overlap integral can be regarded as a Fourier transform
of the wave functions.
CO
<
CO
O
rr
Q.
(5
UJ
z
Fig.7 : (a) Measured dl/dVj, curves in the temperature Hg.8: Calculated tunneling probability and the
range between 1.7 K and 40 K. (b): dI/dVb curves of measured dI/dVb curve at 1.7 K. The downward
unstructured samples arrows mark the resonance peaks. The upward
arrows mark the k=0 positions for each subband.
168
Fig.8 shows a comparison between the calculated tunneling probability and the
dl/dVb curve measured at 1.7 K. As one can see, the experimental results agree well with
the calculated peak positions. The lowest three OD subbands are determined as AE01-7
meV, AE12=6meV,AE23=5 meV. The relation between the width of the potential, w,
and the wave vector, k , w=ip /k, be used to check the order of magnitude of w for the
lowest OD subbands. The upward arrows mark the position, where k=0 for all OD
subbands. and thus, the voltage spacing DV can be used to determine k using
eAV=AE=h2k2/2m* which then yields a value of w=94 nm, 102 nm, and 120 nm for
the nA= 1,2,3 subbands. These values are in good agreement with the geometrical
diameter of the dot of dgeo=160 nm if the surface depletion is taken into account.
5. Conclusion
References
University of Toronto,
Dept. of Electrical & Computer Engineering, 2Dept. of Chemistry and
ia
Laser & Lightwave Research Centre; Toronto, M5SI A4 CANADA
1. Introduction
It has been the "proven truth" for some decades that the performance of electronic
devices should rise with decreasing dimensions. The continuos advance of lithographic
technology in the past made possible serious exploration of the fascinating world of
low-dimensional structures and was in turn propelled by the results of this exploration.
Today, the submicron lithography in combination with different growth techniques
such as MBE and MOCVD is routinely used for fabrication of quantum-well based
devices with superior characteristics.
The situation is changing however as researchers move toward one- and zero-
dimensional structures. Although there have been a number of successful efforts, e.g.
fabrication of free-standing Si or Se nano-wells or nano-pillars by the less controllable
thinning techniques after conventional lithography or e-beam lithography, the now
celebrity lithographic technology seems to approach its resolution limits here, making
the cost of controllable manufacturing of high-density arrays of 1-D and 0-D structures
almost prohibitively high for industrial implementation as the dimensions of individual
units fall to 10 -100 nm.
The situation has forced us to look for new approaches for the nano-manufacturing.
One way is to use different methods of direct writing of desirable structures on
semiconductor or dielectric wafers, e.g., the focused ion beam technology or the atomic
force microscope (AFM) and perhaps STM. The method looks promising for relatively
small structures, but becomes unrealistic when a large area or a large number of
devices are needed. One alternative is to make use of spontaneous island formation in
planar growth of some lattice-mismatched semiconductor systems, that gives more or
less disordered 2-D arrays of nano-sized clusters [1]. It is not yet clear however
whether the usefulness and cost efficiency of this method are sufficiently high to
outweigh the. considerable sacrifice in the ability to control the structure parameters in
this case.
171
5. Luryi etal. (eds.), Future Trends in Microelectronics, 171-183.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
172
Another very different approach is non-subtractive
direct fabrication by selective deposition in the
openings of some sort of template. This of course
could be a regular lithographic structure, with all the
above mentioned limitations.
On the other hand, one can use different self-
organised regular nano-structured (nano-porous)
templates, such as zeolites and molecular sieves [2],
polymer nuclear trucks membranes [3] or porous
anodic aluminum oxide (AAO) films [4]. This is a
particularly attractive approach as it circumvents the At substrate
The preparation procedure for porous AAO films with different pore diameters is
described in [4]. AC electrolysis was used for the deposition of metals (Ni, Co, Fe and
others) and compound semiconductors (CdS, CdSe, CdSxSei.x and GaAs) resulting in
the formation of nano-wire arrays in the pores of anodic alumina films. Deposition of
other semiconductors is the subject of our current efforts. Electrodeposition was
performed at room temperature for metals and GaAs; and at 100 - 160°C for the
^IlgVI systems. Metals and GaAs deposition was carried out from aqueous acidic
electrolytes, whereas, AHB^ were deposited from solutions containing metal salt and
elemental chalcogen in dimethylsulphoxide [17]. Although different in execution, the
technique is similar in scope to the one used to deposit 200 nm diameter cadmium
selenide/telluride microwires in the pores of Anopore membranes [18]. However, the
large pore diameter of most available Anopore films precludes observation of low-
dimensional effects.
Take the fabrication of CdS nano-wires as an example. Electron microphotographs
(fig. 2) show that the deposited material fill the pores uniformly. Quantitative electron
microprobe analysis indicates a 1:1 stoichiometry to within the detection tolerance of
the technique. XRD data of the AACACdS nano-wires have diffraction patterns
corresponding to the hexagonal. The unit cell of AACACdS was found to have
smaller parameters and greater density than what is reported for the bulk. Annealing
at 500°C for 1 hour successfully relieved the lattice distortion except for samples with
dp<12 nm.
The X-ray diffraction data suggest that the spatial confinement of the deposition
affects the crystallite orientation. The relative intensity of the (002) diffraction peak of
the AACACdS is considerably higher than those recorded for the Pt\CdS (DC
deposited) sample or a crystalline "bulk-like" CdS powder therefore indicating that
crystallites growth with the c-axis oriented along the pore. Annealing only slightly
enhances the diffraction patterns of AACACdS, preserving its existing (002) texture.
The contributions of other primary diffraction planes ( (100), (101), (110) ) are
suppressed significantly. Yet another template-induced structural feature of the
AACACdS is that crystal domains size changed very little after annealing, suggesting
that most of the deposited CdS is crystalline and the CdS domains span the widths of
the pores. We are tempted to conclude that the main factor determining the observed
174
Figure 2. (a.) SEM micrographs of CdS nano-wires partially exposed by dissolution of the oxide film; (b.)
TEM micrographs of the CdS nano-wires after complete dissolution of AAO supporting matrix; (c.) TEM
micrograph of a microtomed cross section of the Al\AAO\CdS film.
structural features of the AAO\CdS, and especially its texture, is the spatial
confinement inside the pores.
mB
D
~~X 2.37
~—"*JI
c> A
2.35
A
V/ l_^A-_. _ 10 15 20 25
250 500 750 1000 1250
1 diameter / nm
Raman shift / cm
Figure 3. Resonance Raman spectra of AAO\CdS Figure 4. Band gap as a function of CdS nano-wires
nanowires arrays with light polarized across the wires diameter. Symbols - experiment, line - calculation.
fundamental mode decreases with decreasing wire diameter, suggesting that the
quantum size effect regime has been approached (Fig. 3). Based on this dependence we
calculated band gap energies and other parameters by fitting experimental data to the
expression which describes the ratios of the overtone intensities as a function of
excitation energy [20,21].
The band gaps determined with polarisation across the wires ranged from 2.376 eV
for the large pore diameters to 2.417 eV for the smallest pore diameter and are well
behaved as a function of the nano-wire diameter (Fig. 4) determined by the expression
for spherical particles [22] when bulk values are used for the effective masses of the
electron and hole, and the optical dielectric constant; 2.41 eV is assumed for the band-
gap energy of bulk CdS. The calculated overtone intensity ratios values for ss-
polarisation are consistent with the experimental data (Fig. 4).
0.2
455 465 475 485 495 505
excitation wavelength, nm 10 20 30 40 50 60 70
aspect ratio/1 d-1
Figure 5. RRS overtone ratio as a function of Figure 6. Measured perpendicular and parallel
excitation wavelength for different CdS wire diameter. coercivities of AAO\Fe wires as a function of wire
Symbols - experiment, lines - calculation. aspect ratio. Squares - HCJ_, circles - Hc||.
176
3.2. MAGNETIC EFFECTS IN NANOWIRES.
Iron nano-wire arrays with varying diameters were fabricated [11] and their magnetic
properties determined. The coercivity was found to be highly anisotropic and
dependent on the aspect ratio of the particles (Fig. 6). The functional dependence of Hc
on the aspect ratio suggested that the metal deposit consists of a cylindrical assembly of
fused single-domain particles.
60
:
40
20
Figure 9. Two-wire double-junction system. Arrows show the prevailing directions of tunneling for V>0.
Using the "global rule" of SET [27] to calculate the effect of wire charging on the
tunnelling probabilities, we performed Monte-Carlo simulations of the electron
transport through the system assuming a low-impedance environment. Typical
numerical values used in modelling were: junction capacitances Cr = (1.6±.2) -10"19 F,
coupling capacitance C0 =(0 to 200)-Q, junction resistances Ru = Rl2 = 50kQ
(source), R2}=200-Ru and R^ = (1 to 100) • i^, (drain). We assumed zero
conductance between wires.
4.1 RESULTS
Analysis that uses the "global rule" ideology shows that in the case of very strong
coupling (C0 »Cr) the whole array seems to respond as a single double-junction
system wherever the electron actually tunnels, and the relative contribution of the
particular wire charge to the system "charging energy" is proportional to a small ratio
CriC0. The apparent consequence is the reduction of the Coulomb-blockade region in
proportion to the number of participating wires. Attempt to apply this finding directly
178
to our experimental system is impeded by present lack of knowledge of what specific
wires are participating in the charge transport at any given moment, knowing a finite
dispersion in oxide thickness and a large dispersion injunction resistances.
The interesting part is that strong coupling also leads to spontaneous polarisation of
neighbouring wires, when an accumulation of excessive electrons on one wire is partly
compensated by the hole accumulation on another one, yielding very low-energetic
"excitonic excitation" of the system. The polarisation is conveniently described by the
difference in charges on individual wires P = N2-Nl (in units of electron charge e). It
can considerably exceed the total charge N = Nl+N2 on both wires, which at low
enough temperature remains rather strictly determined by the applied voltage.
50
|v=2V |
40
Polansalion
^30 ■
a»
°20
o
10 ■
Noise
,
10
Figure 10. Average polarization and polarization dispersion vs. drain resistance asymmetry; T=0, Q IQ ^ IQO •
For identical wires P stochastically oscillates in time around zero average value
(P), with increasing with coupling dispersion DP=J((AP)2). The polarisation
becomes however quite regular in the case of asymmetrical wires, when electrons can
tunnel, say, from one of the wires "faster" than from the other. Figure 10 illustrates
1.5 2.0
Voltage (V)
Fig. 11. Average wire charges (N.\ VS. voltage at three different temperatures. Top 3 curves represent the
"slower" wire, other 3 - the "faster" one. "A" and "V" show the "faster" and "slower" wire charges at a particular
time instant for T=0 K
179
rapid increase of (P) with increasing asymmetry of the drain resistances, with a
maximum saturation level in the order of C0/(Cn +C21) (determined by the "faster"
wire).
The effect is voltage- and temperature-sensitive. It practically disappears at the
critical voltages corresponding to steps on the I-V curves. At these points the
maximum total charge on both wires N increases by one. This brings about a new
phenomenon which can be called the Coulomb blockade of polarisation. The wire
charges, (P) and Dp thus periodically oscillate with increasing voltage (Fig. 11).
The strong system polarisation in-between these critical voltages is a SET-induced
effect which does not exist for high temperatures when kT > e2 llC, and the thermal
fluctuations control the charge statistics. For higher temperatures the average wire
charges approach the Kirchhoff s values at all voltages and vary almost linearly with
the voltage (500K- curves in Fig. 11). For lower temperatures the polarisation statistics
is governed instead by the shot noise in combination with the Coulomb-blockade effect,
which suppresses the fluctuations of the total charge on both wires but increases the
anticorrelated fluctuations of the individual wire charges.
This behaviour is illustrated in Fig. 12 which shows how (P) and DP depend on
temperature for "peak polarisation" (left graph) and "low polarisation" (right graph)
voltages. In the first case the almost constant average polarisation is accompanied by
the increasing noise as temperature increases. What is interesting and counter-
intuitive is that not only the average polarisation, but also the polarisation noise
decrease at first as the temperature rises, and only at higher temperatures the
polarisation noise start rising again due to the thermal contribution.
1 1 I 1 ' ' 1 '
V =2.2V v=2V DP
J
*,2Q
« <p> "
=r~-\
O
Df - <p>
-
. . 1 . . . i .
200 400 200 400
Temperature, K Temperature, K
Figure 12. Average polarization and polarization noise vs. temperature for different voltages.
x 10'
T= 0 K
■^ 10"
| 10"'
o
"5 10"'
I to-1
10"'
10'
Figure 13. Spectral density of fluctuations for the wire charge polarisation and the total wire charge at two
temperatures.
and that of the charge polarisation P much more striking - at low temperatures
2 2
(AP )(O) = 0) exceeds /AW )0 = 0) by 5 orders of magnitude.
Thus, under the SET conditions the wire coupling leads to strong interwire
polarisation of charges, which for just two wires stochastically oscillates in time and
periodically - as function of the applied voltage. In the much more complex case of
multi-wire systems and nanowire arrays, this SET-induced spontaneous self-
polarisation may lead to a number of (quasi)static and dynamic phenomena. Among
others, the prospect of "phase transition" between random polarisation and ordering
with changing of system parameters or an applied field, and the possibility of the self-
sustained "polarisation waves" are being investigated.
6. References
1. Leon, R., Petroff, P.M., Leonard, D. and Fafard, S. (1995) Spatially resolved visible luminescence of
self-assembled semiconductor quantum dots, Science 267, 1966-1968.
2. Ozin, G. A. (1992) Nanochemistry: synthesis in diminishing dimensions, Adv. Mat. 4, 612-648.
3. Martin, C. R. (1994) Nanomaterials: a membrane based synthetic approach, Science 266,1961-1966.
4. Al-Mawlawi, D., Liu, C. Z. and Moskovits, M. (1994) Nanowires formed in anodic oxide
nanotemplates, J. Mater. Res. 9, 1014-1018; Al-Mawlawi, D., Douketis, C, Bigioni T. et al (1995)
Electrochemical Fabrication of Metal and Semiconductor Nano-wire Arrays, Proc. 187 Meet.
Electrochem. Soc, May 21-26, Reno, Nevada, Pennington, in print.
182
5. O'Sullivan, J.P. and Wood, G. C. (1970) The morphology and mechanism of formation of porous anodic
films on aluminium, Proc. Roy. Soc. Land. A. 317,511-543.
6. Switzer, J. A., Hung C. J., Breyfogl B. E. et al (1994) Electrodeposited defect chemistry superlattices,
Science 264,1573-1576.
7. Ross, C. A. (1994) Electrodeposited multilayer thin films, Annu. Rev. Mater. Sei. 24,159-188.
8. Streltsov, E.N. Osipovitch, N.P. Routkevitch, D. L. and Sviridov, V.V. (1995) Electrochemical
deposition of compositionally modulated bismuth chalkogenide films, in preparation.
9. Gregory, B. W. and Stickney, J. L. (1991) Electrochemical atomic layer epitaxy (ECALE), J.
Electroanal. Chem. 300,543-561.
10. Kawai, S. and Ueda, R. (1975) Magnetic properties of anodic oxide coatings on aluminum containing
electrodeposited Co and Co-Ni,J. Electrochem. Soc. 122,32-36; Shiraki, M., Wakui, Y. and Tsuya, N.
(1985) Perpendicular magnetic media by anodic oxidation method and their recording characteristics,
IEEE Trans. Magn. 21, 1465-1467;
11. Al-Mawlawi, D., Coombs, N. and Moskovits, M. (1991) Magnetic properties of Fe deposited into anodic
aluminum oxide pores as a function of particle size, J. Appl. Phys., 70 4421-4425; Dunlop, D. J., Xu,
S., Özdemir, O., Al-Mawlawi, D. and Moskovits, M. (1993) Magnetic properties of oriented iron
particles as a function of particle size, shape and spacing, Phys. Earth Planet. Inter., 76, 113.
12. Haruma, M., Kobayashi, T., Sano, H. and Yamada, N. (1987) Chem. Soc. Jap. Chem. Lett., 407;
Miller, D. and Moskovits, M. (1989) Separate pathways for the synthesis of oxygenates and
hydrocarbons in Fisher-Tropsch reaction, J. Am. Chem. Soc, 111, 9250.
13. Foulke, D. G and Stoddard, W. B. (1963) in F. A Lowenhiem (ed.) Modern Electroplating, Wiley,
New York, 632; Preston, C. K. and Moskovits, M. (1988) New Technique for the determination of
metal particle size in supported metal catalyst, J. Phys. Chem., 92, 2957-2960, ; Preston, C. K. and
Moskovits, M. (1993) Optical characterization of anodic aluminum oxide films containing
electrochemically deposited metal particles, J. Phys. Chem. 97, 8495-8503; Saito, M., Kirihara, M.,
Taniguchi, T. and Miyagi, M. (1989) Micropolarizer made of the anodized alumina film, Appl. Phys.
Lett., 55,607.
14. Pontifex, G H., Zhang, P., Wang, Z., Haslett, T. L., Al-Mawlawi, D. and Moskovits, M. (1991) STM
imaging of the surface of small metal particles formed in anodic oxide pores, J. Phys. Chem. 95, 9989-
9993.
15. Whitney, T. M., Jiang, J. S., Searson, P. C. and Chien, C.L. (1993) Fabrication and magnetic properties
of arrays of metallic nanowires, Science 261,1316;
16. Nagodawithana, K., Searson, P.C., Liu, K. and Chien, C.L. (1995) Processing and properties of
electrodeposited Cu-Co multilayered nanowires, Proc. 187 Meet. Electrochem. Soc, May 21-26, Reno,
Nevada, Pennington, in print.
17. Baranski, A S. and Fawcett, W. B. (1980) The electrodeposition of metal chalcogenides, J.
Electrochem. Soc. 127,766-767.
18. Klein, J. D., Herrick, R. D., Palmer, II, D., Sailor, M. J., Brumlik, C. J. and Martin, C. R. (1993)
Electrochemical fabrication of cadmium chalcogenide microdiode arrays, Chem. Mater., 5,902-904.
19. Brus, L. E. (1984) Electron-electron and electron-hole interactions in small semiconductor crystallites:
the size dependence of the lowest exited electronic state, J. Chem. Phys. 80,4403;
20. Routkevitch, D., Ryan, L. and Moskovits, M. (1995) in preparation.
21. Sliiang, J. J., Risbud, S.H. and Alivisatos, AP. (1993) Resonance Raman studies of the ground and
lowest electronic excited state in CdS nanocrystals, J. Chem. Phys. 98, 8432.
22. Wang, Y. and Herron, N. (1991) Nanometer-sized semiconductor clusters: materials synthesis, quantum
size effects, and photophysical properties, J. Phys. Chem. 95,525-532; Sweeny, M. andXu, J.M. (1989)
Hole energy levels in zero-dimensional quantum balls, Solid State Comm., 72, 301.
23. Al-Mawlawi, D., Moskovits, M., Ellis, D., Williams A and Xu, J.M. (1993) Working with Mother
Nature: Nano-wire diode arrays made by novel techniques and functioning at 300K, Proc. of 1993 Int.
Device Research Symposium, Virginia, USA, 311.
24. Haruyama, J., Moskovits, M., Routkevich, D., Tager, AA and Xu, J.M. (1995) Periodic conductance
oscillations in novel nano-wire/semiconductor anti-dot arrays, Submitted to the IEDM'95 Conference.
25. Mullen, K., Ben-Jacob, E., Jaklevic, R.C. and Schuss, Z. (1988) I-V characteristics of coupled
ultrasmall-capacitance normal tunnel junctions, Phys.Rev.B 37, 98-105.
183
26. Mooij, J.E. and Schön, G. (1992) Single charges in 2-dimensional junction arrays, in H.Grabet and
M.H.Devoret (eds.), Single Charge Tunnelling, Coulomb Blockade Phenomena in Nanostructures,
NATO ASI Ser. B 294, Plenum Press, New York., pp.275-310.
27. Averin, D.V. and Likharev, K.K. (1991) in B. Altshuler, P. A Lee, and R. A Webb (eds.), Mesoscopic
Phenomena in Solids, Elsevier, Amsterdam, Chapt.6.
28. Misuki, I., Yamamoto, Y., Yoshino, T. and Baba, N. (1987)7. Met. Finish. Soc. Jap. 38, 561.
29. Chou, S.Y., Wei, M.S., Krauss, P.R. and Fisher, P.B. (1994) Single-domain magnetic pillar array of 35
nm diameter and 65 Gbits/in2 density for ultrahigh density quantum magnetic storage, J. Appl. Phys. 76,
6673-6675.
30. Piraux, L., George, J.M., Despres, J.F., Leroy, C. et al (1994) Giant magnetoresistance in magnetic
multilayered nanowires, Appl. Phys. Lett. 65, 2484-2486; BlondeL A, Meier J.P., Doudin, B. and
Ansermet, J.-Ph. (1994) Giant magnetoresistance of nanowires of multilayers, Appl Phys. Lett. 65,
3019-3021.
31. Masuda, H. and Fukuda, K. (1995) Ordered Metal Nanohole Arrays Made by a Two-Step Replication
of Honeycomb Structures of Anodic Alumina, Science 268, 1466-1468.
32. Moskovits, M. and Xu, J. M. (1994) Nano-electric devices, USA Patent application.
TAMING TUNNELLING EN ROUTE TO MASTERING MESOSCOPICS
Abstract:
A single tunnel barrier in an asymmetric doping environment exhibits some
superior properties as a microwave detector diode, when compared with any competing
device. The manufacturability of even this simple structure is not yet established. One
cannot yet procure semiconductor epitaxial material that routinely meets the tight
tolerances on the specification required to produce diodes with a sufficiently small spread
of de I-V and microwave characteristics. Wafer-scale uniformity and wafer-to-wafer
reproducibility are key issues, as are the techniques used for qualifying wafers. Proposals
for devices that exploit mesoscopic systems will appear all the more realistic if tunnelling
devices are established as manufacturable.
1. Introduction:
device is that the current-limiting mechanism is tunnelling through a barrier rather than
thermionic emission over a barrier. As a direct consequence, the sensitivity to ambient
temperature of the microwave properties is greatly reduced [5]. It has a wider dynamic
range than the Ge backward diode which relies on interband tunnelling, and is still used
because of the temperature independence of its performance. In low-cost communication
systems, the ability to dispense with temperature compensation circuits is attractive. The
temperature-dependence that remains comes from residual thermionic emission over the
barrier.
A simulation package has been developed for help in the design of the device [5].
188
This incorporates a self-consistent solution of the Poisson equation in the n-i-n doping
structure (including an accumulation layer on the cathode side of the tunnel barrier), and
uses the Thomas-Fermi approximation to obtain a starting potential profile. The current
density passing through the tunnel barrier is then calculated by standard tunnelling theory
techniques. There are some further important issues of detail, such as the determination of
the Fermi energy throughout the structure. While this is easy at zero bias, the way Ef
varies between the two contacts by an amount eV when a bias V is added is non-trivial.
It is approximated here by being assumed constant though the undoped regions of GaAs
from the appropriate contacts, and taking an average value within the barrier. No account
is taken of tunnelling through indirect gap states in the AlAs or any inelastic processes
during tunnelling or roughness of the tunnel barrier profile.
In designing to a specification, the incompleteness of the simulation package
itself is a problem. If devices that are claimed to be grown within specification give
different I-V characteristics, how much of the difference is due to the inadequate model and
how much to variances between grown structure and the specification? We return to this
problem below. In Figure 2, we show the tolerancing on our diode structure by virtue of
monolayer errors in the barrier thickness and shortfalls in aluminium fraction in the
barrier. The sensitivity to these variances is indicative of the exponential sensitivity of
the current to the width and height of the barrier. This sensitivity has been the major
hurdle in establishing manufacturable tunnel devices. Are cryrtal growers working in the
commercial, as opposed to research, sphere able to meet specifications, such as shown in
Table 1 below, with sufficient yield to provide diodes costing 20(2, as required for
automobile radar at 76GHz or ozone monitoring at 170GHz? Multiple regrowths to hit
the target, or the need to test individual devices in the event of device-to-device
fluctuations, simply rule out this device as uncompetitive. There are other sensitivities in
the design associated with the level of doping in the contact layers, and the length of the
shorter undoped GaAs layer, but these are less critical than the details of the barrier layer.
There is an important issue of metrology that runs in parallel with the design and
growth to a tunnel device specification. There are no wafer-scale, non-destructive,
methods for qualifying wafers bought in from a commercial supplier. Furthermore, all
conventional materials assessment methods, photoluminescence (PL), transmission
electron microscopy (TEM), secondary ion mass spectroscopy (SIMS), and even X-ray
diffraction (XRD) fall short in one or more ways in being able to confirm whether an as-
189
grown wafer is within specification [12]. Whereas TEM can specify chemical
composition of layers with good detail, the data is valid only over a small region of the
wafer surface (of linear dimension -0.1 (im), and nothing can be inferred about the doping
profile. SIMS allows the doping and composition profiles to be registered with respect to
each other, but not with the resolution demanded by the tolerancing. Both techniques are
destructive and unsuitable for wafer-mapping. XRD can map out the uniformity of the
GaAs layers above the AlAs, but it does not routinely have the resolution to determine
that the total amount of Al in the AlAs layer is both laterally uniform and within the
specified composition profile.
Wafers in batches of five have been ordered from commercial suppliers of MBE
and MOCVD epitaxial multilayers, against the specification given in Table 1. Of the 13
suppliers approached only half quoted, the others not being able or willing to meet either
the uniformity target or even the specification itself. (The MBE wafers were grown
sequentially, while the MOCVD layers were grown simultaneously in a multi-wafer
machine.) Representative data taken from MOCVD grown wafers by TEM, SIMS and
XRD is collated and shown in Figure 3 as a radial map. The TEM gives an accurate
thickness of the AlAs at some point on the wafer, and its depth from the top surface.
Lateral variations, including the distribution of steps between layers of monolayer
thickness differences, can also be determined on a length scale of <0.1|J.m. With the help
of detailed simulations of the electron diffraction patterns, the aluminium composition
profile through the barrier can also be determined. The SIMS allows the dopant (Si)
distribution to be determined, and its general registry with respect to the Al concentration.
Here there is a specific problem: the mass of A1H is the same as that of Si28, and
sputtered Al which picks up hydrogen from the background contaminates the Si signal,
preventing an accurate registry of the doping profile with respect to the Al profile, and
obliterating part of the Si profile. A more demanding Si29 SIMS analysis is required.
The depth resolution of SIMS is insufficient in some cases even to detect some of the
thinner layers in our structures. Absolute calibration of SIMS concentrations remains
difficult, and in any event only the dopant concentration, not the doping concentration, is
measured. The information about the size and shape of the Al profile (and a fortiori its
wafer-scale uniformity) is contained in the relative amplitude of the successive peaks in
the large angle X-ray data, and the signal-to-noise in a standard instrument is found to be
insufficient - a full triple axis spectrometer may be required. The 'noise' in the high angle
X-ray data (from which the Al profile would be reconstructed) may be intrinsic to the high
doping level in the contact layers in our sample. The differences in thicknesses of layers
5-7 inferred by the different techniques is noted, and needs resolution.
Viewed as a radial map, the uniformity of the thicknesses of the layers seems
satisfactory. It appears that the GaAs layers above the AlAs layer, and the AlAs layer
itself are both slightly thinner than specified and there may even be systematic variations
in the thicknesses as interpreted by the various techniques.
190
1.0 Layers
, —i—i—i—i—i—i—i—i—i—■—■—■ ' i
5-7: GaAs above AlAs bamer
'
Black=XRD
I #4 Shaded=SIMS ...j
E 0.9 +10%
Open=TEM
I#5""
% 0.8 W ÜF
5. -10%
I 0.7
0.6 20.0 25.0
-5.0 0.0 5.0 10.0 15.0
Radial position (mm)
4.00 ■ r i ■ i ■< I ' i ■ .
: AlAs barrier (TEM data) •#4
■#5
1 3-50
1 3.00 - - +10%
^_
■
•■ -10%
£ 2.50
:
, i . i . i 1 , 1
2.00
0 10 5 15 20 25
Radial position (mm)
Figure 3: The radial uniformity of the MOCVD samples as implied by TEM, SIMS and
XRD data (a) the GaAs layers above the AlAs barrier layer and (b) the AlAs barrier layer
itself. It appears that all layers are systematically thinner than the target specificaton.
-i ' > i 1 " ' ' 1 ' ' ' ' M ' _ In Figure 4, we show the range of de I-V
2.0
: MOCVD #1 !l : characteristics taken from another MOCVD-
grown wafer in the same batch. Here the
picture is potentially more hopeful. It
Z Wafer centre/ :y I
appears that the uniformity of the I-V
■4
1.0 1 (4 devices) Jl _
characteristics as averaged over 60|J.m
ü
; Haiffuir / \ ; diameter mesas from wafer #1 is quite
< - radius jl \ adequate for production (a 6% spread from
: (4 devices) J '.
centre to mid-way out, and 16% from centre
0.0 ^^Xwafer edgq
to edge). A 4Q. series resistance gives
- ^1<^ (5 devices) -
's/s?' - quantitative agreement between simulation
and experimental I-V characteristics. The
i n results from wafer #2 (with 20^m diameter
-1.0 -0.5 0.0 0.5 1.0 diodes) seem quite unsatisfactory, with a
100% spread in the I-V data without a
Voltage (V)
systematic trend, and the surface appearance
Figure 4: The radial variation in dc I-V suggests that this may be due to process-
characteristics taken from an MOCVD induced artifacts. Results on wafer #3 are
wafer. similar to those shown in Figure 4.
191
The performance of this type of diode as a microwave detector has already been
established [5]. As a part of this exercise, the uniformity of the rf parameters will be
correlated with the dc parameters. Accelerated life testing will be undertaken: the active
region of the device is single crystal material, so the device is likely to be robust [11].
The MBE wafers were grown sequentially. The same type of characterisation
techniques have been applied, and the structural results inferred from TEM, XRD and
SIMS analysis of two wafers are shown in Figure 5. All three techniques indicate a radial
variation of the thickness (of layers 5-7) not present in the MOCVD data, and this has
been seen in previous exercises [12]. Again there is a significant difference between the
results inferred from the different analytical techniques. The SIMS data indicates that the
doping level within the contacts is within specification, although the doping in layer 2 is
not. The thickness of the AlAs barrier is within specification, but slightly thicker than
the target figure. (The MOCVD layers were thinner than the target figure, and this
difference has been noted before [5].) There is no electrical data as yet.
1.0 -i—I—I—I—n—r 111
i '—i ■ i | i -|—i—i—i—i—|—i—i—i—r-
Layers 5-7: GaAs
C above AlAs barrier Black=XRD
? 0.9 •GB1355 Shaded=SIMS .
'GB135H 0"peri=TEM +10%
8 0.8 Q.
5. ILT
-®. ®
j -10%
i 0.7
0.6
-5.0 0.0 5.0 10.0 15.0 20.0 25.0
Radial position (mm)
4.00 : r
" l ■ i ' i
; AlAs barrier (TEM data) ' ' #GB1355
■GB1358 _
3.50 tr
+10%
| 3.00 •■ •■ '■
-
-10%
: i . i
, i , i . i
2.00
0 5 10 15 20 25
Radial position (mm)
Figure 5: The radial uniformity of the MBE samples as implied by TEM, SIMS
and/or XRD data (a) the GaAs layers above the AlAs barrier layer and (b) the AlAs barrier
layer itself.
The ability to fabricate very thin layers with adequate precision and uniformity
for tunnel-based microwave diodes would be encouraging for the feasibility of using
mesoscopic systems as elements of future computational systems. Our 20|i.m diameter
microwave diodes may still be averaging the properties of the tunnel barriers, or even
exploiting regions of thinnest and lowest Al composition. Even so, the modest level of
variability (and the systematic variation across the wafer) of the diode device performance
is encouraging.
If one continues to retain the conventional methods for device and circuit
fabrication based on epitaxy and lithography at the level of mesoscopic devices (with
active volumes and contact areas characterised by 10-100nm sides), then a comparable
exercise involving electron-beam lithography is needed. The resistances of individual
members of a series of wires of sub-0.1u.m diameter of ~10|J.m length will have to show a
small standard deviation, whether those wires be in metals, suicides or doped
semiconductors.
The much tougher exercise of integrating both technologies at the mesoscopic
level has started in a limited number of physics-based contexts. Starting with the work
of Reed et al [13], a number of exercises have sought one-dimensional resonant tunnelling
in -O.ljim diameter pillars. Tewordt et al [14] have produced results with an enormous
variability in I-V characteristics between diodes of comparable feature size. Indeed, some
workers [15] have used the variability to map out the strategic locations of donor levels
that play a controlling role in the level of current that a diode can transmit. It is a matter
of urgency that an exercise comparable to that reported in section 2 is undertaken to
ascertain the uniformity of I-V characteristics that can be achieved, and their stability under
accelerated life-tests. A major difficulty here is the absence of any self-limiting process in
operation while etching is performed: if such a process could be invented, it would have a
major impact.
193
4. Summary
5. Acknowledgements
6. References
R. A. SURIS
A. F. Ioffe Physical-Technical Institute
194021 Polytechnicheskaya 24, St. Petersburg, Russia
1. Introduction
fi2q2 fi2(q-Q.)
*M"
2TI
M M>k|eiQl'Zk)|2(NQ+l) h2k-
2m + 2m +^0
F
■-h a,
4rcee* /zN
vQn = 2MO) 0
VQZ+Q? V Q
where e* and M are effective charge and ion reduced mass, ©Q is longitudinal
phonon frequencies (the phonon wavenumber Q = (Q& Qi), Q. is the total system
volume. \|/k is electron wave function of the continuous spectrum, v|y0 and E0 are the
wave function and ionization energy of the localized state and
NQ=(exp(toQ/T)-l) •
Here e0 and sw are the dielectric constant values at low and high frequencies.
Averaging Sk with Boltzman distribution function gives
60-sm e2 ft© T
S-TC--^—-•—-• — • — • N„ + l
e„ fo E■'on En \ Q
1
For the sake of simplicity we shall not consider the effect of resonant states on the capture process.
2
Naturally, Hits perturbation theory result is valid if Sk, q is less than the velocity in the z-direction, ftk/2m.
3
By the way, the parameter Gx does not depend on the distances between QWs because vj CC 1/ L while
xocL.
200
A the characteristic scale of lifetime is 10"6s. This is due to overlapping the energy
spectra of free electrons and of electrons localized in QWs in the z-direction.
The situation can be
radically improved by
replacing of QWs by QDs. In 10l ■;^(H
Ü
this case, we have a real ü«\vS<oS«w»*»K;
Figure 4
201
trivial: the similar situation is ordinary for atoms and here we deal with artificial
atoms. However, there are two important distinctions in the case of semiconductor
with QDs.
The first one is a conceivable possibility to make these artificial atoms
have the desirable spectrum. It can be done by choosing the appropriate sizes of
QDs and band offsets.
The second remarkable and very important distinction consists in the
possibility of dc current pumping inherent in semiconductor heterolasers. This is a
current of consequent electron tunneling between the neighboring QDs that is
providing by tunable thickness and height of the potential barrier between QDs.
Let us try to imagine some thinkable schemes of cascade lasers with QDs.
The first one uses four-level system (Figure 5). The electric field is applied
Figure 5 Figure 6
to the semiconductor with 3D array of QDs. When the field provides the resonance
between the ground electron state of n-th QD and the first excited states of the
neighboring (n+l)-th QD, their energy levels are splitted into two ones (see Figure
5). If the splitting value is close to the phonon energy and temperature is low, the
filling factor of upper splitted level is less than filling factor of the lower one.
Therefore, we have the inverse occupation in the system of hybridized states of the
neighboring pairs of QDs (Figure 5). This picture is valid while the splitting value
is large as compared with /i/xr where xx is the characteristic time of electron
transitions between the splitted levels. It implies a large value of overlapping
integrals of wave functions of neighboring QDs and low temperature.
Other scheme uses three-levels in each QW (Figure 6). This scheme would
work when the rate of electron transitions 2 =>1 inside a QD exceeds the rate of the
tunnel transitions 1 => 3 between the neighboring QDs. In this case we have the
inversion between the levels 3 and 2 inside each QD.
eNT(bopt) ,
•nun
Jth — —^ ^- + VeboptBnlPl ebop,BnlP] I+
I T
QD \bop,ry(bopt)s
where the dimensionless parameter s is the ratio of the stimulated transition rate in
QDs at the lasing threshold to the spontaneous transition rate in the narrow-gap
region
Bn,p,
The optimum thickness of the OCL, bop'\ and the optimum surface density of QDs,
NgPt, are, in their turns, the functions of the dimensionless parameter s. By
way of illustration we considered the following double-heterostructure laser
structure: the materials of wide-gap regions (cladding layers), narrow-gap region
(OCL) and QDs are InP, Gao.21Ino.79Aso.45Po.54 and Gao.47Ino.53As respectively, the
latter two being lattice-matched to InP. For die total
204
1
By way of illustration we 1
considered the following double- 1* 1
1
■B (b)
1
heterostructure laser structure: the ■*->
1
£• 1 sS
materials of wide-gap regions (cladding **> 'M
<S £?
1
I
y^s
sy
layers), narrow-gap region (OCL) and §0 1 y^
1
QDs are InP, Gao.21Ino.79Aso.4ePo.54 and 22 ö
C 0
O H
1
1 //
/y
eN ebB
fnfp +
L
QD~ v„vp^D(l-fn)(l-fp):
where fn,>p are the filling factors (common to all QDs in this case) satisfying the
threshold condition <f^> + <f>-i = NIT / n,p n.p'n.p' (J are the cross
sections of electron and hole capture into a QD, vnp are the thermal velocities of
electrons and holes. Here the ratio of the stimulated transition rate in QDs at the
lasing threshold to the spontaneous transition rate in the narrow-gap region,
s = (\/fy(a/ry)(F/xph)(Ae).ilmm/Bnlpl, plays the role of a universal
dimensionless parameter controlling the magnitudes of the minimum threshold
current density, optimum surface QD density and the optimum thickness of the
OCL.
205
£2 BO.QO —
"c:
32 e~J,-
.40.00 —
5 cr
1 cr
O.OO 0.04- O.OB 0.12 0.16 0.20 0.24-
a 4-0 cr
o
■>».
rvj
<n 1
E
O
0 mm 20 cm
Cl> O
o
5_ 2.00- = 10 cm
-1
ß
ß = 5 cm-
ß = 1 cm
o 1
'' ' ' ' I'''
0.20 D.24
Figure 9
Here Q= eFa/A is the Bloch oscillation (BO) frequency, n is the electron density
and the brackets {..) mean averaging over initial wavenumber distribution.
The BO can manifest themselves as microwave and far IR radiation.
However, the BO are very fast smoothed by electron scattering processes and after
the relaxation the voltage dependence of current is described by the N-shaped
characteristics (Ref.8):
Aa / , \ nx~
j = e—-n-^coskj-
» ° l+("x«)
4 . . . 2
This result can be easily obtained from more general expression for eigenfiinctions presented in Ref
5
Here for the sake of simplicity we suppose the QDs forming a rectangular lattice with periods ax, ay, az.
In principle, when the rations of Fx, Fy, and ?z are rational there are sets of different Nx, Ny and Nz values
giving the same energy value. In this case the resonant transitions are possible. Here we do not consider this
case and restrict ourselves to the following remark. If all of this rations are far from unity the probabilities of
these transitions are exponentially small.
208
semiclassical approach we obtain the following equations for the current density
components
ja=e^-n-(cos(koaaj)-sin(Qat), Qa=eFaaa/ft, a = x,y,z
Therefore, the current projection on the certain direction is the sum of three
harmonics with the frequencies that can be tuned by applied field rotation.
One can expect that this feature should significantly broaden functionality
of the BO applications.
5. References
1
R. Kazarinov and R. Suris, Theory of electrical properties of semiconductors with
superlattices, Sov. Phys.- Semiconductors, 1973, 7, no. 3, p. 347
2
R. Kazarinov and R. Suris, Possibility of amplification of electromagnetic waves
in a semiconductor with superlattice, Sov. Phys.- Semiconductors, 1971, 5, no. 4, p.
707; Electric and electromagnetic properties of semiconductors with a superlattice,
Sov. Phys.- Semiconductors, 1972, 6, no. 6, p. 120,
3
J. Faist, F. Capasso, D. L. Sivco, C. Sirtori, A.L. Hatchinson and A.Y. Clio,
Quantum cascade laser, Science 1994, 264, 553; Electron. Lett. 1994, 30, 865
4
Y. Arakawa, H. Sakaki, Multidimensional quantum well laser and temperature
dependence of its threshold current, Appl. Phys. Lett., 40, 939, (1982).
5
N. Kirstaedter, N. Ledentsov, M. Grundmann, D. Bimberg, V. Ustinov, S.
Ruvimov, M. Maximov, P. Kop'ev, Zh. Alferov, U. Richter, P. Werner, U. Gölsele,
J. Heydenreich. Electron. Lett., 30, 1416 (1994). Low threshold, large T0 injection
laser emission from (InGa)As quantum dots, Electron. Letters, 30, 1416, (1994).
6
R. Suris and L. Asryan, "Quantum-Dot Laser: Gain Spectrum Inhomogeneous
Broadening and Threshold Current", Proceedings of SPIE's 1995 International
Symposium on Optoelectronic, Microphotonic & Laser Technologies.
PHOTONICS WEST'95, 4-10 February 1995. San Jose, California USA, v. 2399,
pp.433-444.
L. Asryan and R. Suris, "Linewidth Broadening and Threshold Current Density of
Quantum-Box Laser", Proceedings of International Symposium Nanostructures:
Physics and Technology. June 20-24, 1994, St.Petersburg, Russia, pp. 181-184.
7
R. Suris and S. Shtofich, Role of impurities in the appearance of multifrequency
emission from injection semiconductor lasers, Sov. Phys.- Semiconductors(July
1983) 17, no 7, 859
8
L. Esaki and R. Tsu, Superlattice and Negative differential conductivity, IBM J.
Res. andDev. 14,1970, p 61
ARCHITECTURES FOR NANO-SCALED DEVICES
LEX A. AKERS
Centerfor Solid State Electronics Research
Arizona State University
Tempe, AZ 85278-5706
1. Introduction
The desire for higher performance, low cost electronic systems seem insatiable. We
have witnessed in the latter part of the 20th century the introduction of microcomputers
and cellular telecommunications systems. These systems have undergone continued
enhancements in computational power, memory, and special features. The demand for
further system improvements, lower power consumption, anywhere-anytime access to
data and communications, multimedia, and portable personalized digital assistants will
continue[l]. Also many real-time applications such as vision and speech recognition,
robotics, and numerous other interactive control and signal processing applications will
require hundreds of gigaflops of processing speed[2]. The scaling of device feature sizes
into nanometer dimensions can conceivably allow systems made with these components
to fulfill the performance improvements desired.
The advantages of scaling devices to smaller sizes are overwhelming since the area
per function, the energy needed to switch a device, and the energy needed to be stored to
represent information scale with the physical size of the device. However, the common
assumption that improvements in device performance will automatically translate into
improved system performance is not true. We believe one major challenge to
implementing such systems using nano-scaled devices is finding appropriate system
architectures to host these devices. The modularity and locality of the architecture and
the length of the connections between devices and between subsystems determine if the
improvement in device performance predicted by scaling theory is reflected in
improvement in system performance. A discussion of these issues is presented,
followed by recommendations on the types of architectures that allow designers to better
capitalize on the characteristics of nano-scaled devices.
2. Architectural Issues
10000 _
9000 --
8000 ..
7000 --
6000 ..
5000 --
c
4000 ..
3000 --
2000 --
1000 ..
0 + +
0.35 0.25 0.18 0.13 0.1 0.07
Minimum Dimension(um)
—Ave connection
capacitance(ff)
—Total device
Cap(2Cg+Cd)(ff )
30.00
25.00 --
- Connection
Power
dissipation
10%(W)
-I/O Power
Dissipation
10%(W)
the predicted I/O needs in conventional architectures^]. This difference is shown in Fig.
4. These architectures provide system performance that will improve with device
performance while keeping the overall power dissipation, shown in Fig. 5, down to
acceptable levels.
5.00E+03 -T-
4.50E+03
4.00E+03
3.50E+03
1.50E+03 ..
l.OOE+03
5.00E+02
0.00E+00
Minimum Dimension(um)
Figure 4. I/O for traditional and locally interconnected architectures vs. minimum
feature size
4. Conclusion
Scaling will allow devices to shrink to nanometer dimensions and result in vast
improvements in their performance. However, to translate this improvement into
system improvements will require system architectures which use local rather than
global information and communication for computations.
References
7. Akers, L.A., Walker, M., Ferry, D.K., and Grondin, R., (1989) A Limited-Interconnect,
Highly Layered Synthetic Neural Architecture, VLSI for Artificial Intelligence, Kluwer
Academic Press.
8. Matsuzawa, A. (1994) Low-Voltage and Low-Power Circuit Design for Mixed
Analog/Digital Systems in Portable Equipment, IEEE J. Solid-State Circuits 29, 4, 470-
480.
9. Ferry, D.K., Akers, L.A., and Greeneich, E., (1988) Ultra Large Scale Integrated
Microelectronics, Prentice-Hall.
70.00 _
60.00
50.00 ..
.x\ -■ Total Power
dissipation 10%
40.00 ..
(W)
■o
-D Total Local
30.00 --
20.00
X Architecture(10
%) Pd(W)
10.00 ..
0.00 + + + +
0.35 0.25 0.18 0.13 0.1 0.07
Minimum Dimension(um)
1. Introduction
Simulation of semiconductor devices has reached some maturity for device
structures that can be described by the system of device equations given
by Shockley [1, 2]. High energy transport, including hot electron effects
such as impact ionization and gate currents, can be correctly simulated by
full band Monte Carlo approaches [3] (solving Boltzmann-type equations)
and has also matured into the realm of engineering; the missing pieces
being mainly standardization and numerical efficiency. Correspondingly,
commercial packages, which solve the Shockley equation system (even in
three dimensions) and feature full band Monte Carlo post processors, are
available or are in the final stages of development. Complex full band Monte
Carlo device codes are also available [4].
The treatment of (abrupt) heterojunctions in devices has not yet ma-
tured to the desirable degree and is almost certainly required to be un-
derstood in the future in great detail. One can make a case, and current
developments point strongly to it, that heterolayers must more and more
replace the dilute donor/acceptor doping configurations as the sizes of de-
vices decrease toward the typical donor/acceptor spacing. The simulation of
abrupt heterostructures invariably involves quantum mechanics and solu-
tions of the Schrödinger equation. The quasi-two-dimensional electron gas
of the metal-oxide semiconductor transistor is very well investigated and
understood [5] and work on quantum wires and dots is in progress.
Complex problems are involved, however, in the coupling of these quan-
tum regions to the classical "Shockley regions". This coupling needs to be
215
S. Luryi et al. (eds.), Future Trends in Microelectronics, 215-225.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
216
accomplished on various levels. On the physical side much progress has been
made with Landauer-Biittiker type of coupling which in many instances is a
generalization (and specialization with regard to dimensionality) of Bethe's
thermionic emission theory. From a numerical point of view the quantum
region necessitates multigrid approaches, where in the simpler cases the
grid describing the quantum region is typically reduced in dimension (by
one) compared to the reservoirs. In this way one can construct a multiscale
approach that combines classical and quantum regions in one simulation
entity. We have performed such a task for semiconductor quantum well
lasers diodes and will report this as an example below. While these simu-
lations deal mainly with current and resistance concepts, the quantization
also influences the concept of capacitance and inductance and examples
of atomistic features of quantum dot capacitance have been given [6]. The
knowledge of electronic band structures also becomes increasingly impor-
tant as shown by full band Monte Carlo simulations, calculations of optical
transitions in lasers, density of states calculations, etc.
The next step in scaling down, however, involves the mesoscale and
molecular scale and requires still more elaborate quantum mechanical meth-
ods to obtain insight into the electronic structure. One might, for example
have in mind a quantum dot containing only a very limited number of con-
stituting atoms, a tunneling tip, etc. At this scale the dynamics of atoms
also assumes a special importance as can be seen from experiments offeree
and tunneling microscopes that move and switch atoms. The method of
Car and Parrinello [7] emerges as a powerful tool to simulate this scale and
in turn can feed back to the bigger scales giving information about surface
and interface structure and dynamics.
A complete coupling of these methods to understand questions on all
scales is still in the future. However, some coupling of those methods has al-
ready been achieved and representative examples are given in the following
sections.
Figure 1. Schematic conduction band edge for a quantum well laser diode
These are approximately classical 3-D regions, and transport can be mod-
eled with the Shockley equations. Carriers move ballistically over the abrupt
heterojunctions and thermalize in roughly a mean free path. Since the quan-
tum well (QW) can be on the order of the mean free path or much smaller,
injected carriers may traverse the well, reflect, resonate above the well, or
inelastically scatter into bound states. The coupling of this quantum region
to the classically behaving diode is a complicated mesoscopic problem that
must be treated by a Bethe-Landauer type of approach. Bulk drift-diffusion
is coupled in our simulation [8] to ballistic injection into 3D (continuum)
states above the well. Continuum carriers can then transfer to 2D "bound
states" through a net capture rate given by [9]:
4 6
Frequency (Hz) x10
(a) 1 (b)
Jr"1 ..-■■■•"'T"---..
§ J,MI.....---- ; I JL,
Figure S. Probability density obtained by SEMC for the quantum well system of Figure
1 in the (a) normally incident and nominally resonant trunk state and (b) corresponding
branch (scattered) states with strong inelastic scattering. Ex. is the component of energy
associated with motion normal to the plane of the quantum well. The well is located
between 20 nm and 30 nm.
fa)
&0
Z
1
w
Q
03
11
O
os
(d)
Figure 4- (a) Probability density in the nominally resonant incident electron state as a
function of position and imaginary absorbing potential —ihR/2, where iJis the scattering
rate, and (b) transmission, (c) reflection and (d) transmission probabilities as a function
of incident energy and scattering rate R. The quantum well is located between 20 nm
and 30 nm in (a).
3. Atomic-Scale Dynamics
As feature sizes approach the atomic scale, we can no longer rely on the
effective mass approximation and other such continuum approximations.
Rather, ions are modeled by pseudopotentials that determine their elec-
tronic orbitals. Thus, with present levels of computing power, this down-
222
ward scaling restricts us to simulation of only relatively small systems. In
most cases, however, this restriction is not prohibitive since only a limited
number of atoms need be considered in such atomic-scale systems. Further,
at this scale, defects and disorder can be treated in a straightforward man-
ner, and dynamic effects related to the motion of the ions, such as their
thermal vibration, can be studied at the most basic level.
In our simulations, electrons are described by density functional theory,
and the motion of the ions obeys Newton's equations with forces given by
the gradient of the total energy of the system. Both the electrostatic repul-
sion between the ions and the contribution from the electronic clouds are
included. In this way, the many-body effects for the electrons are greatly
simplified. This method has been proven to be extremely reliable for calcula-
tion of ground state properties such as the total energy and the equilibrium
configuration of the ions. For properties, such as optical absorption and
transport, that involve excited electronic states, extensions of the method,
e.g. the Generalized Gradient Approximation, greatly improve the density
functional theory.
Figure 5. Total electronic density for the tip-surface system. The crosses indicate the
positions of the silicon atoms.
mizing the total energy of the system. Various sophisticated methods have
been devised for solving both equations [20]. We are currently concentrat-
ing on a minimization approach that appears to be quite natural to density
functional theory. We have tested several minimization methods including
the linear search algorithm and the conjugate gradient method, and we
have considered the functional form for the total energy that requires ex-
plicit orthonormalization of the electronic orbitals as well as the form that
avoids this constraint. We have found that a modification of the simple
linear search procedure for the functional form requiring explicit orthonor-
malization is the most efficient algorithm for the systems of moderate size
that we have studied so far. In this method, part of the total energy gradient
is integrated exactly over one iteration step. However, for larger systems
we expect that the functional for non-orthonormal orbitals will be more
efficient.
To study the dynamic properties, we use the method introduced by R.
Car and M. Parrinello [7]. They have shown that it is possible to write
equations of motion for the coupled system of the electrons and the ions
such that the electrons oscillate around their instantaneous ground state
while the ions obey Newton's equations of motion. In other words, the sys-
tem oscillates close to the Born-Oppenheimer surface [19], and it is not
necessary to compute the electronic ground state for each ionic configura-
tion. This method is very efficient, and a vast literature has demonstrated
its applicability to a broad range of problems.
Combined, this set of first-principles tools has been tested for bulk mate-
rials, and excellent agreement is routinely achieved for the equilibrium lat-
tice constant, the bulk modulus, and the bulk phonon frequencies computed
both within the frozen phonon approximation and with molecular dynam-
ics. We also obtain a reconstruction of the clean and of the H-passivated
silicon (001)-(2xl) surface in agreement with recent highly converged re-
sults in the literature.
As an example of such atomic-scale calculations, the interaction of a
STM tip with a H-passivated silicon surface has been simulated. The sur-
face is modeled as a slab of eight layers of silicon atoms with the atoms in
the four upper layers relaxed to their equilibrium positions. The STM tip is
represented by a single silicon atom adsorbed to the back of the slab. Figure
5 shows the distribution of the total electronic density before the thermo-
stat is switched on. The hydrogen passivation induces a reconstruction of
the surface with the formation of symmetric dimers in the [110] direction.
When the thermostat is switched on (the bath is set at room tempera-
ture), the atoms start to oscillate, and Figure 6 shows the corresponding
variation of the tunneling barrier height between the tip atom and the sur-
face. This fluctuation is larger than the thermal energy spreading, and a
224
-0.05
£-0.06
§< -0.07
ft -0.08
U
& -0.09 h
m
-0.1
Figure 6. Fluctuation of the top of the tunneling barrier due to the thermal motion of
the tip and surface atoms.
Acknowledgement
References
1. Streetman, B.G. (1980) Solid State Electronic Devices, Prentice-Hall, Englewood
Cliffs, NJ.
2. Hess, K. (1988) Advanced Theory of Semiconductor Devices, Prentice-Hall, Engle-
wood Cliffs, NJ.
3. Bude, J. (1991) Scattering Mechanisms for Semiconductor Transport Calculations,
225
Monte Carlo Device Simulation: Full Band and Beyond, ed. K. Hess, Kluwer Aca-
demic Publishers, Norwell, Mass.
4. Laux, S.E. and Fischetti, M.V. (1988) Numerical Aspects and Implementation of the
DAMOCLES Monte Carlo Device Simulation Program, Monte Carlo Device Simu-
lation: Full Band and Beyond, ed. K. Hess, Kluwer Academic Publishers, Norwell,
Mass.
5. Ando, T., Fowler, A.B., and Stern, F. (1982) Electronic Properties of Two-
Dimensional Systems, Review of Modern Physics, 54, 466.
6. Macucci, M., Hess, K., and Iafrate, G.J. (1995) Simulation of Electronic Properties
and Capacitance of Quantum Dots, /. Appl. Phys., 77, 3267-3276.
7. Car, R., and Parrinello, M. (1985) Unified Approach for Molecular Dynamics and
Density-Functional Theory, Phys. Rev. Lett. 55, 2471-2474.
8. Grupen, M., Ravaioli, U., Galick, A., Hess, K., and Kerkhoven, T., (1994) Cou-
pling the Electronic and Optical Problems in Semiconductor Quantum Well Laser
Simulations, Proc. SPIE OE/LASE Conf., 2146, Los Angeles, CA, 133-147.
9. Grupen, M., Kosinovsky, G., and Hess, K. (1993) The Effect of Carrier Capture on
the Modulation Bandwidth of Quantum Well Lasers, 1993 International Electron
Device Meeting Technical Digest, 23.6.1-23.6.4.
10. Grupen, M. and Hess, K. (1994) Self-Consistent Calculation of the Modulation
Response for Quantum Well Laser Diodes, Appl. Phys. Lett. 65, 2454-2456.
11. Nagarajan, R., Mirin, R.P., Reynolds, T.E., and Bowers, J.E. (1993) Experimental
Evidence for Hole Transport Limited Intensity Modulation Response in Quantum
Well Lasers, Electron. Lett. 29, 1688-1690.
12. Shichijo, H., Kolbas, R. M., Holonyak, N., Dupuis, R. D., and Dapkus, P. D. (1978)
Carrier Collection in a Semiconductor Quantum Well, Solid State Communications
27, 1029.
13. Brum, J. A. and Bastard, G. (1986) Resonant Carrier Capture by Semiconductor
Quantum Wells, Phys. Rev. B 33, 1420-1423.
14. Sotirelis, P. and Hess, K. (1994) Electron Capture in GaAs Quantum Wells, Phys.
Rev. B 49, 7543-7547.
15. Preisel, M. (1994) Carrier Capture and Carrier Kinetics in Biased Quantum Well
Devices, Tele Danmark Research, H0rsholm, Denmark.
16. Register, L.F. and Hess, K. (1994) Numerical Simulation of Electron Transport in
Mesoscopic Structures with Weak Dissipation, Phys. Rev. B 49, 1900-1906.
17. Hess, K., Register, L.F., and Macucci, M. (1994) Toward a Standard Model in
Nanostructure Transport Problems Including Dissipation, Proceedings of the 2nd
International Symposium on Quantum Confinement: Physics and Applications 94-
17, 3-17.
18. Zory, P.S. (1993) Quantum Well Lasers, Academic Press, San Diego.
19. Pastore, G., Smargiassi, E., and Buda, F. (1991) Theory of Ab Initio Molecular-
Dynamics Calculations, Phys. Rev. A 44, 6334-6347.
20. Payne, M.C., Teter, M.P., Allan, D.C., Arias, T.A., and Joannopoulos, J.D. (1992)
Iterative Minimization Techniques for Ab Initio Total-Energy Calculations: Molec-
ular Dynamics and Conjugate Gradients, Rev. of Mod. Phys. 64, 1045-1097.
MONTE CARLO SIMULATION FOR RELIABILITY PHYSICS MODELING
AND PREDICTION OF SCALED (100 NM) SILICON MOSFET DEVICES
1. Introduction
Since the early 1970's, silicon integrated circuit technology has been propelled by
continual and successful efforts to reduce the active channel length of MOSFET devices
[1 - 3]. This exercise in scaling provides the framework that has produced increases in
the density of devices on a chip, increases in device frequency response and operating
speed, and increases in the precision required to achieve more complex systems with
greater functionality and performance. Today, devices with channel lengths well below
100 nm have been produced in many research laboratories, and the downward scaling
trends of the past twenty years are expected to persist at the same pace until at least the
40 nm generation in manufacturing [1], or about the year 2015. This level of technology
is expected to correspond to 128 GBit DRAMs and 28 Ggate microprocessors with five
times the clock frequency and 1/9 the power consumption of today's devices, all
operating at a power supply voltage of around 0.5 V.
Historically, silicon MOSFETs designed for increased performance through scaling
and scaling-related approaches to drain and channel engineering have also produced
compromises in device reliability. It was anticipated that as the effective channel length
was scaled below the electron mean free path (Leff < about 200 nm), non-local transport
effects (such as velocity overshoot and quasi-ballistic transport) would actually enhance
device performance. Furthermore, by scaling the power supply voltage below the
Si/Si02 interface barrier height (i.e., qVrjD < 0B = 3-1 eV), it was expected that many
device reliability problems could be suppressed or eliminated. A new regime for device
operation was predicted where electron energy distribution functions would be "cooler"
than those deduced from models based on the local electric field. However, drain
current degradation was observed in recent experiments on 150 nm floating gate
MOSFETs biased with a drain voltage as low as 1.5V, and no discontinuities in
degradation were observed as the voltage was reduced below 3.1 V [4]. These
experiments confirmed that the degradation was due to non-tunneling current, via carrier
injection over the interface barrier and into the oxide. For such low bias, the channel
electrons cannot gain enough energy from the drain electric field alone to surmount the
barrier; thus, those rare electrons that do inject must experience some type of "energy
227
S. Luryi et al. (eds.), Future Trends in Microelectronics, 227-236.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
228
gaining" interactions to obtain the required energy. Physical mechanisms that produce
this effect are not fully understood, but Monte Carlo simulations indicate that electron-
electron interactions near the drain edge play a significant role [6]. For 100 nm devices
biased at low voltages, high-energy phenomena such as injection into the oxide are
extremely rare and their effects are difficult to measure experimentally. Therefore,
attempts to understand fundamental limits of silicon technology for next generation
devices will need to rely heavily on accurate simulation and modeling methods that can:
(a) identify those phenomena that occur at high energies; (b) predict the consequences of
design and processing on device performance, and (c) analyze device and circuit
reliability and operational lifetime.
Figure 1 illustrates the general technology and design environment for down scaling of
MOSFET devices. This schematic shows relationships between simulation, modeling,
and reliability evaluation/prediction (right side) to basic elements of the scaling process,
manifested through process technology, device structure choices, and experimental
evaluation (leftside). In this context, it is important to recognize the many challenges
for process technology, device
structure choices, and experimental PROCESS
methodology required to fabricate PROCESSING SIMULATION
TECHNOLOGY AND
100 nm MOSFETs [7]. For MODELING
example, use of highly-controlled i i
future experimentation and new physical interpretation [2, 9]. Fortunately, the
concurrent rapid development of computer technology will allow this high degree of
simulation and modeling of reliability physics to be implemented on a personal
workstation in a computing and network environment for integrated circuit design that
has become known as technology computer-aided design (TCAD) [2,8].
Figure 2 shows a schematic for elements of an initial attempt to implement the TCAD
design environment discussed in Section 2. The goal of this approach is to simulate the
"hot electron problem" in small silicon
MOSFETs at dimensions of 100 nm and
SSUPREM4 below. Hot electron phenomena in the
Process channel is the focal point for most device
Simulator
level reliability issues. Important elements
i ,
SPISCES
of the hot electron problem include: (a)
electron heating by transport through the
Device channel; (b) highly-localized lateral and
Degradation Simulator
vertical electric field distributions between
1 ' gate and drain regions; (c) electron
Monte Carlo injection into the SiC>2 gate insulator with
Device a nominal injection barrier of 3.1 eV; (d)
Electron Simulator
Phenomena hot-carrier-induced generation and local
1 ,* charging of Si/SiC>2 interface states; (e)
Oxide
k.
electron transport in the gate insulator; (f)
M<>dei degradation of the gate insulator during
injection and transport; (g) modification of
the electrostatic potential and channel
mobility by charges in the oxide and at the
Figure 2. MOSFET reliability physics interface; (h) induced changes in device
simulation and modeling environment. characteristics, such as threshold voltage,
subthreshold slope, drain current drive, transconductance, etc.; and (i) imminent
expiration of device lifetime and circuit failure due to device characteristic changes.
The approach shown in Fig. 2 accounts for most of these effects, except for transport in
the oxide. Recent reports have shown that this effect can be considered using drift and
diffusion models in TCAD approaches for large devices [7]; however, our focus here is
on the role of Monte Carlo simulation in the development of accurate reliability models
for very small, next generation devices.
As shown in Fig. 2, SSUPREM4 is used as the process simulator to design two-
dimensional device structures. Data from SSUPREM4 (e.g., realistic doping profiles,
process-dependent physical parameters, etc.) are provided as input for the device
simulator, SPISCES, which provides standard electrical characteristics and device
parameters for later determination of device degradation. These device parameters are
also incorporated as initial estimates (e.g., initial electric field distributions, carrier
distributions, etc.) for the Monte Carlo simulator. The Monte Carlo method is used to
(particularly) model the high-energy tails in the electron energy distribution and the
resultant hot electron effects [6,13].
230
The Monte Carlo method [6] employs a realistic silicon band structure for the two
lowest conduction bands that are calculated from the pseudopotential method. A 2-D
solver for Poisson's equation is coupled to the Monte Carlo method, and the solver is
rapidly updated with a time step of At = 0.1 fsec, providing a dynamic electric field
distribution for calculating long-range Coulomb interactions. The scattering routines
include mechanisms for acoustic and intervalley phonons, ionized impurities, impact
ionization, interface scattering, and electron-electron (short-range Coulomb)
interactions. An enhanced particle statistics algorithm generates an ensemble of about
30,000 superparticles to provide necessary details for the high energy component of the
electron energy distribution function. This algorithm is an essential feature to the
overall success of the model.
Prediction of hot-electron-induced oxide damage is performed by adapting Yasuda's
emperical model for interface state generation [6,14]. This uniform carrier injection
model is modified to account for conditions of non-uniform injection near the drain that
are typical for MOSFET devices. Model parameters are determined from data obtained
from charge pumping experiments, and the validity of the approach was demonstrated
for a 1000 nm device [6]. This approach is utilized to obtain distributions of interface
states for 100 nm devices since, to our knowledge, experimental techniques are not able
to resolve the spatial distribution of interface states for small MOSFETs. The resulting
induced device degradation is determined by incorporating these interface states into the
SPISCES simulator (Fig. 2). We assume that interface states are uniformly distributed
in energy throughout the bandgap, and that they are donor-like below midgap and
acceptor-like above midgap. The location of the quasi-Fermi levels at the interface
determines the fraction of filled states [15]. Our approach, including Yasuda's model,
has recently received some additional affirmation from simulations of pMOSFETs [7].
A variety of device structures have been studied using the approach described in Sect. 3.
These include conventional n-MOSFETs, drain- and channel-engineered n-MOSFETs,
and SOI n-MOSFETs. The effects of constant field scaling and generalized scaling on
hot electron phenomena have been investigated for conventional devices with channel
lengths from 1000 nm down to 100 nm and applied source-to-drain voltages as low as
4V As the
VDS = 1- - applied voltage was reduced below the Si/Si02 barrier height, no
conditions were observed for which abrupt changes in slope or discontinuities occurred
in simulated measures of hot carrier phenomena. This is consistent with recent
experimental observations [5]. Thus, charge injection into the oxide is observed for
voltages less than half the barrier height. Our simulations indicate that the lateral
electric field may be increased with each scaling generation for an equivalent rate of hot
electron injection. For applied voltages as low as 1.5V and channel lengths as low as
100 nm, our simulations show that lucky electron models and average electron
temperature models are inaccurate for predicting the total rate and spatial distribution of
hot electron effects in this regime of device scaling. Monte Carlo simulations
demonstrate that, due to electron-electron interactions near the drain edge, energy
distributions of hot electrons entering the drain of small devices do not cool as rapidly as
the average electron temperature nor as rapidly as the electric field decays. Thus, it
seems that hydrodynamic and/or energy balance models will have to be modified (if
231
Figure 3. Energy w
.-£ 10
distrib. functions
for Si MOSFETs.
Curves cl,c2, bl,
b2 use constant
field scaling; al
and a2 use
generalized
scaling [6]. The
applied voltages
are as shown.
Dashed curves
are without 2 3
electron-electron Energy (eV)
scattering; solid
curves include electron-electron events.
These issues are further exemplified in Figure 4, which demonstrates the effects of
high-energy carrier transport in a 100 nm MOSFET under low-voltage bias conditions
(VDS=2VGS=1-5V). Fig. 4 displays the (normalized) lateral electric field, the average
electron energy, and the distribution of electron injection for the channel region between
source and drain. The physical channel is located from 160 nm to 260 nm, and most of
the spatial variations in the quantities of interest occur in the drain region. First, it is
observed that the average electron energy is spatially retarded from the electric field,
with the peak in average energy occurring 5 nm "ahead" of the peak electric field.
Furthermore, the peak in electron injection is also retarded another 15 nm ahead of the
peak in average electron energy—a location where the average energy is well below its
peak value. Hydrodynamic and/or energy balance models will not predict such
retardation effects. The predicted total spatial retardation of the injection is 20 nm-a
significant 20% of the 100 nm channel length. Monte Carlo scaling analysis has shown
that this retardation effect is independent of channel length; thus, retardation will
become an increasing fraction of the channel length as devices are scaled below 100 nm.
Thus, the resultant hot electron and spatial retardation effects in electron injection will
most likely become very important in future scaled MOSFETs.
Figure 5 shows electron distribution functions calculated from the Monte Carlo
simulations at positions A, B, C, and D in the source-drain region and along the channel
232
a n-MOSFET using conventional processing techniques and other devices based on more
novel concepts for drain- and channel-engineering [16]. Table I lists the device
parameters for an array of five structures used in this comparative reliability study. The
first two columns are for drain-engineered structures and the last two columns are for
channel-engineered structures. The conventional structure is used as a baseline device.
Device Structures
Junction 35 30 50 46 30
Depth (nm)
Source/Drain Large
Extension Dopant 7xl018 degenerate degenerate degen-
Doping (cm'3) (Sb) and In erate
Halo
Retrograde
Channel 6.5 xlO 17
6.5 xlO 17
6.5 xlO 17
from 1016 to 6.5 xlO17
Doping (cm"3) (uniform) (uniform) (uniform) 1018in30 (uniform)
nm
Threshold
Voltage (V) at 0.41 0.38 0.40 0.41 0.41
VDs = 0.05V
As shown in Fig. 2, the Monte Carlo simulator is used to calculate high energy
transport properties such as hot electron injection into the oxide for each n-MOSFET
design for VDS=2VGS=1-5V stress conditions. This injection data is then coupled with
oxide damage models to produce distributions of interface states. These interface states
are inserted into the SPISCES device simulator as interface charge determined by the
electron quasi-Fermi level at the interface (see Sect. 3) to evaluate hot-electron-induced
degradation of device characteristics. For example, Figures 6a and 6b illustrate
simulated saturation drain current degradation (AID/ID) versus time for devices listed in
Table 1. Our reliability simulation method predicts some interesting trends for these
designs. First, variations in the drain design have a more pronounced effect on overall
234
10
a
y ■*' ^~^
sO) 10J
y ■*■*' /^
V
a * ••■"*
''.«' /
/
Figure 6. Simulated hot-electron-induced drain current degradation for the array of (a)
drain-engineered and (b) channel-engineered device designs listed in Table I.
device reliability than variations in the channel design. Next, the lighfly-doped-drain
(LDD) provides little electric field reduction over the conventional design; thus, the
LDD experiences comparable electron injection. On the other hand, due to the LDD's
reduced drain doping, it cannot efficiently screen interface charge, and thus suffers
considerably greater degradation than the conventional design. Last, among the
channel-engineered designs, the fully-depleted (FD) SOI design suffers the greatest
drain current degradation. This result can be mainly attributed to interface coupling
effects due to substantial electron injection and subsequent oxide damage in both the
front and back oxides. It is also worthy to note for the array of designs that the hot
electron injection distributions have relative magnitudes that correlate directly to the
relative magnitudes in peak parallel electric field, and the injection distributions are
displaced about 20 nm beyond their respective highly-localized electric field
distributions, as also demonstrated for the 100 nm device in Fig. 4. The results of Fig. 6
predict that the conventional design is the most reliable device in the array of designs
considered. This outcome indicates that although there is a fundamental change in the
underlying mechanisms that produce high-energy carriers as device size and applied
voltage are scaled down (i.e., electron-electron interactions become more significant),
the trade-off between device performance and reliability is still quite significant.
Another point from Fig. 6 is that the drain current does not conform to the model
that predicts a square root dependence of degradation on the stress time [see discussion
in Section 3 of Ref. 6]. The slope of the degradation curves is particularly dependent on
the interface state density and the resultant carrier recombination and trapping effects
that can occur. In particular, in Fig. 6, the FD-SOI device appears to have a smaller
curvature, which could be due to the fact that both oxide interfaces participate in the
degradation process. In an attempt to further evaluate this degradation signature, we
studied two additional 100 nm FD-SOI designs. These include a heavily-doped channel
(SOI-1) and a lightly-doped channel (SOI-2) device design, with a uniform channel
doping of 1 x lO1^ cm"^ and 1 x 101" cm"-5, respectively. The silicon layer thickness,
front oxide thickness, and back oxide thickness are 30 nm, 4 nm, and 80 nm,
235
respectively, for both devices. The gate workfunction is adjusted to produce equal
threshold voltages of 0.4V for both devices. Figure 7 shows the simulated drain current
degradation in both the
linear and saturation
regions of the current-
voltage characteristics of
SOI-1 and SOI-2. The
SOI-2 design experiences
greater degradation
mainly due to increased
coupling of interface
states through the lightly-
doped channel. It is also
observed that these two
devices experience
greater degradation rates
than the 100 nm designs
listed in Table 1. This is
consistent with the
interpretation that the
degradation of drain
current for SOI devices is
proportional to hot
electron injection and Stress Time (seconds)
subsequent oxide damage
in both the front and back Figure 7. Drain current degradation characteristics
oxides. for SOI-1 and SOI-2.
6. Conclusion
A Monte Carlo simulator has been developed and combined with a process simulator, a
device simulator, and oxide damage models to investigate hot electron phenomena and
device reliability for deep submicron (100 nm) n-MOSFETs and SOI devices designed
for low-power applications. Monte Carlo simulations indicate that non-local transport
and two-dimensional effects in the drain current and electric field distributions influence
hot electron injection into the oxide(s) for both bulk MOSFET and SOI devices. Also,
electron-electron interactions play an important role in the creation of the high energy
tail of the electron energy distribution function for devices under low-voltage bias
conditions (qVos< 0ß)- The Monte Carlo results are linked to the reliability simulation
method to investigate hot-electron-induced device degradation. This method predicts
that as devices are scaled down to the 100 nm regime, aggressive drain designs can
provide increased performance while reducing the trade-off of reduced reliability. The
LDD design concept, while improving short channel effects, no longer reduces device
degradation when compared to the conventional design. Finally, ultrathin-film, fully-
depleted SOI designs greatly improve short channel effects; however, they experience
considerably greater device degradation than their bulk MOSFET counterparts due to
hot-carrier-induced damage to both the front and back oxides.
236
7. List of References
1. Hiroshi Iwai, Hisayo Sasaki Momose, Masanobu Saito, Mizuki Ono, and Yasuhiro
Katsumato (1995) The Future of ultra-small geometry MOSFETs beyond 0.1
micron, Microelectronics Engineering 28,147-154.
2. Armin W. Weider (1995) Si-Microelectronics: Technology Perspectives-Risks,
Opportunities, and Challenges (This Volume).
3. G. Baccarani, M. R. Wordeman, and R. H. Dennard (1984) Generalized Scaling
Theory and Its Application to a 1/4 Micrometer MOSFET Design, IEEE Trans.
Electron Devices ED-31,452-459.
4. J. E. Chung, M.-C. Jeng, J. E. Moon, P.-K. Ko, and C. Hu (1990) Low-Voltage
Hot-Electron Currents and Degradation in Deep-Submicrometer MOSFET's, IEEE
Trans. Electron Devices ED-37, 1651-1658.
5. David Esseni, Luca Selmi, Roberto Bez, Enrico Sangiorgi, and Bruno Ricco (1994)
Bias and Temperature Dependence of Gate and Substrate Currents in n-MOSFETs
at Low Drain Voltage, Proceedings of the International Electron Devices Meeting
(EDM,), 307-310.
6. John J. Ellis-Monaghan, K. W. Kim, and Michael A. Littlejohn (1994) A Monte
Carlo study of hot electron injection and interface state generation model for silicon
metal-oxide-semiconductor field-effect transistors, J. Appl. Phys. 75, 5087-5094.
7. A. v. Schwerin and W. Weber (1995) 2-D Simulation of pMOSFET hot-carrier
degradation, Microelectronic Engineering 28, 277-284.
8. M. Rodder, A. Amerasekera, S. Aur, and I. C. Chen (1994) A Study of Design/
Process Depaendence of 0.25 |J.m Gate Length CMOS for Improved Performance
and Reliability, Proceedings of the IEDM, 71-74.
9. Andrea Ghetti, Luca Selmi, Enrico Sangiorgi, Antonio Abramo, and Franco Venturi
(1994) A Combined Transport-Injection Model for Hot-Electron and Hot-Hole
Injection in the Gate Oxide of MOS Structures, Proceedings of the IEDM, 363-366.
10. S. Jallepalli, C.-F. Yeap, S. Krishnamurthy, X. L. Wang, C. M. Maziar, and A. F.
Tasch Jr. (1994) Application of Hierarchical Transport Models for the Study of
Deep Submicron Silicon MOSFETs, VLSI Symposium Proceedings, 91-94.
11. Marco Mastrapasqua and Jeff D. Bude (1995) Electron and Hole Impact Ionization
in Deep Sub-micron MOSFETs, Microelectronic Engineering 28, 293-300.
12. K. Taniguchi, M. Yamaju, K. Sonoda, T. Kunikiyo and C. Hamaguchi (1994)
Monte Carlo Study of Impact Ionization Phenomena in Small Geometry MOSFETs,
Proceedings of the IEDM, 355-358.
13. R. B. Hulfachor, K. W. Kim, M. A. Littlejohn, and C. M. Osburn (1995) Non-Local
Transport and 2-D Effects on Hot Electron Injection in Fully-Depleted 0.1 |im SOI
n-MOSFETs Using Monte Carlo Simulation, Microelectronic Engineering 28, 175-
182.
14. N. Yasuda, H. Nakamura, K. Taniguchi, and C. Hamaguchi (1989) Interface State
Generation Mechanism in N-MOSFETs, Solid State Electronics 32 , 1579-1586.
15. M. V. Fischetti and S. E. Laux (1988) Monte Carlo Analysis of Electron Transport
in Small Semiconductor Devices Including Band-Structure and Space-Charge
Effects, Phys. Rev. B 38,9721-9730.
16. R. B. Hulfachor, K. W. Kim, M. A. Littlejohn, and C. M. Osburn (1995) A Monte
Carlo Study of Drain and Channel Engineering Effects on Hot Electron Injection
and Induced Device Degradation in 0.1 mm n-MOSFETs, Fifty-Third Annual
Device Research Conference (DRC) Digest, 14-15.
SUPERCONDUCTOR-SEMICONDUCTOR DEVICES
HERBERT KROEMER
ECE Department, University of California
Santa Barbara, CA 93106, USA
1. Introduction
It has long been recognized that electronic devices operating at reduced temperatures—
including both semiconductor and superconductor devices—can often offer much
higher performance (by several criteria) than room-temperature devices. But the need
for cooling has greatly retarded their use, and there exists an almost-universal persistent
belief that low-temperature devices just don't have a chance to find significant practical
applications.
My presentation is based on the premise that this belief is a myth, and that the
future of electronics is likely to draw increasingly, within the next decade or two, on
low-temperature devices, at least in applications such as high-performance workstations
and scientific and medical instrumentation, where increasing performance requirements
can justify the additional cost of the cryogenics, which is itself decreasing
However, the performance-to-cost relation is by no means the only issue: No
matter how favorable that relation is, no system engineer is going to fool around in a
"real" commercial system with cryogenics under conditions that resemble those of a
research laboratory. What is absolutely essential is "user-friendly" cryogenics! The
enabling technology for the widespread actual use of cryogenic electronics is likely to
be the increasing availability of small self-contained closed-cycle refrigerators. The
development of the latter (mainly Stirling-cycle machines), originally driven by IR
detector technology, has more recently found increasing use in high-7"c superconductor
applications. It is rapidly approaching the point that we may begin to view such a
refrigerator as just another module inside a piece of electronic equipment, somewhat
analogous to, say, a fancy high-voltage power supply.
237
S. Luryi et al. (eds.), Future Trends in Microelectronics, 237-250.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
238
Suppose I offered you a self-contained box, about 2-3 liters in volume, drawing
less than 100 Watts, and I would provide inside this box a volume of about 100cm
inside which I guarantee a temperature T, of say, 77K, with a cooling capacity of, say
3-4 Watts. Given a reasonable cost, such a box would evidently meet our demand for
user-friendly cryogenics. The above specifications are not fictitious, they are those of
actual hardware about to go into production, interestingly by a company whose
business is in the field of high-rc superconductors, and which has found it necessary to
provide integrated system solution to its customers, solutions that include a user-
transparent cryogenics package (Superconductor Technologies, Santa Barbara, CA).
The principal bottleneck to their more widespread use is their cost, but this is likely
to follow the classical pattern of dramatic cost reduction in the wake of building up
mass production. Furthermore, the specifications are likely to improve with time,
including rapid progress to lower temperatures with time, at least to about 20K, the
practical limit of the Stirling cycle, with slower progress below that.
Anybody invoking this last scenario as a realistic one for the future must address him-
or herself to the fact that a huge effort of precisely this kind was undertaken by IBM
during the 70-s, only to be abandoned in 1983. The failure of this project had a terribly
discouraging effect on the whole field of low-temperature electronics, and anybody re-
considering this approach is in danger of running afoul of Santayana's famous dictum
that "those who do not remember the past are condemned to repeat it."
It has been argued persuasively by Likharev [3] that this failure was due, not to the
need for liquid-helium temperatures, but to two quite unrelated reasons: (a) The use of a
unsuitable non-refractory metallurgy based on lead as a superconductor, which was not
sufficiently stable under thermal cycling. The resulting reliability problems would have
been avoided by using niobium as a superconductor, (b) The use of a logic principle,
employing voltage-state logic, that was basically too imitative of semiconductor logic,
and which had inherent power dissipation limits that negated much of the speed
advantage of Josephson junctions. As Likharev points out, a much more suitable form
of superconducting logic would be one that is based on the unique property of super-
conductors that magnetic flux in superconducting loops is quantized, and which
shuffles single flux quanta rather than shuffling voltage states. Likharev's own
presentation at this workshop reviews the present state
It would constitute a major breakthrough for superconductor-semiconductor
devices if a high-temperature superconductor could be found that is technologically
compatible with existing semiconductors, especially III-V semiconductors. As it stands
now, all the high-rc superconductors are oxides that must either be deposited, or
require a post-deposit anneal, in a high-temperature oxidizing atmosphere that will
simply destroy any of the semiconductors it is in atomic contact with, thereby elimi-
nating barrier-free structures. Current research on high-rc superconductors stresses the
achievement of higher critical temperature, rather than elimination of the need for a
high-temperature oxidizing environment. From the point of view of super-semi devices,
the achievement of semiconductor-compatible materials would be a far more valuable
goal, even if it meant a drastic reduction in critical temperature, say, to 40K.
240
2. Semiconductor-Coupled Superconducting Weak Links
2.1. INTRODUCTION
Like Josephson tunnel junctions, weak links exhibit a pronounced Josephson effect,
manifested by a current-voltage characteristic as in Figure 2, which shows data from a
semiconductor-coupled weak link of the kind shown in Figure 1. The characteristic
feature of the Josephson effect is the existence of a current range inside which a
resistance-less supercurrent can flow between the two superconducting banks, up to a
241
certain critical current Ic. Only when this current is exceeded does a voltage appear
between the superconducting terminals.
Compared to tunnel junctions, weak links have a much larger inter-electrode
separation between the two superconducting banks, which leads two potential major
advantages: (a) much lower capacitances, an important consideration for the use of
these devices as high-speed devices (b) a much smaller sensitivity of the characteristics
to variations in the electrode separation.
2.0 —I—l—i—l—r-
•2.9 K
■3.9 K
1.0 L = 0.6(im
b=50nm
1 0.0 U
-1.0 .
-2.0 L J i i i—1_ J I I I l_
-2.0 -1.0 0.0 1.0 2.0
V [mV]
The Pair Wave Function and its Phase. The essence of superconductivity is the
existence of a common pair wave function for the Cooper pairs in the superconductor,
which may be written
Here, the magnitude | y/ (r)[of the pair wave function is related to the local Cooper pair
density n(r) via
|VA(r)|2=n(r), (2)
and 0 (r) is a phase. The key point is that this phase is coherent over macroscopic
distances, and, in the absence of a current, it is the same throughout the entire
superconductor.
7 = Ic-sin(02-0i). (3)
Here, with the ordering of the two phases as given, a positive current designates a flow
of Cooper pairs from bank #2 to bank #1. Because the pairs carry a negative charge
-2e, the electrical current is in the opposite direction, from bank #1 to bank #2. In weak
links, more complicated relations may occur, but /(02 - 0i) is always an odd function
243
of the phase difference, and inasmuch as phase differences have a physical meaning
only modulo lit, the I(92- 9^) relation is necessarily a periodic one, with a period 2n.
A.C. Josephson Effect. In the absence of a bias voltage between the two supercon-
ducting banks, whatever phase difference 92 - 9X may be present, will not change with
time, hence the current will continue to flow—which is why it is called a supercurrent.
If an external bias voltage is present, the difference becomes time-dependent according
to the simple law
f(ö2-ö1)=^(y2-y1). (4)
The supercurrent-vs.-phase relation I(92 - 9X) remains valid in the presence of such a
voltage, but the supercurrent now oscillates about zero, with the Josephson frequency
vj=y(v2-y1), (5)
Incident
Electron
#— EiE^b^l-ö—J
kT
Andreev
Hole
(a) (b)
Figure 3. Andreev reflections, (a): Basic concept, (b): Persistent current flow by
multiple Andreev reflections.
In a semiconductor with a large mean free path for the electrons, the Andreev hole
left behind has a large mean free path itself, roughly equal to that of the original elec-
tron, and theory shows that the hole travels back into the semiconductor along a
trajectory that essentially re-traces the trajectory of the original incident electron. If its
mean free path is sufficiently large, the hole will eventually reach the opposite super-
conducting electrode. In the absence of any bias across the structure, the energy of the
hole is still within the superconducting gap on that side. Such a hole cannot enter the
superconductor, but it can be annihilated by breaking up a Cooper pair inside the other
superconductor: One of the electrons of the pair annihilates the hole, the other electron
takes up the annihilation energy, and is injected into the semiconductor as a ballistic
electron above the Fermi level, at an energy exactly equal to that of the initial electron.
This process, illustrated in Figure 3b, can evidently be repeated: The Andreev reflec-
tions act as what I like to call a "Cooper pair pump," annihilating Cooper pairs on one
side, and re-creating them on the other. If all reflections of electrons and holes were
Andreev reflections rather than "ordinary" reflections, and if there were no other kinds
of scattering processes, the result would be a persistent current.
However, perturbations are always present, and we would expect simply an
enhancement of the conductivity by a factor equal to the number of ballistic traverses
before an unfavorable reflection or collision event randomize either the electron or the
hole flow in this chain reaction. Also, even if no unfavorable reflection and collision
events took place, the diagrams in Figure 3b show only half the story: For each state
with a given direction of arrows there exists another state with all current flows
reversed. If both states of such a pair are occupied, their currents will cancel. To
understand how a supercurrent can arise, we must go beyond the pure ballistic particle
246
picture of Figure 3b, and must take into account the wave properties of the unpaired
electrons and holes, and of the Cooper pairs [13].
For every value of n, there will actually be two states, corresponding to opposite
directions of the flow arrows in Figure 3b.
Up to a point, the above is exactly the same condition as for the bound states in an
"ordinary" one-dimensional semiconductor quantum well. These, too, are states for
which the round-trip phase changes are the different multiples of 2%. In fact, with
regard to the spatial confinement of the unpaired electrons and holes inside the semi-
conductor portion of the structure, the stationary states may indeed be viewed as a new
kind of bound states [13], the difference being that the "heterojunction" barriers are
now formed, not by the conventional energy gap of another semiconductor, but by the
superconducting energy gap of the two superconducting electrodes.
However, there are two decisive differences. The first is that for an AR state con-
fined by superconducting energy gap barriers, the phase on one of the two traverses is
carried by an electron, on the other traverse by a hole. This means that these kinds of
bound states actually carry a current across the semiconductor, in contrast to the
current-less conventional bound states in a conventional quantum well. The two states
belonging to a given n belong to opposite directions of that current flow.
A second difference is the following. As in a conventional quantum well with
barriers of finite height, the round-trip phase shift contains a contribution from the
reflections at the two superconductor barriers. In a semiconductor quantum well, these
contributions simply represent the finite penetration of the wave function into the
barrier, and they are responsible for lowering the bound state energies with decreasing
barrier height.
But in the case of Andreev reflections there is an additional phase shift at each
bank, equal in magnitude to the phase of Cooper pair wave function in that bank, but
with a sign depending on whether an electron or a hole is reflected: When an electron is
reflected at a superconductor with phase 0, the wave function of the hole resulting from
the reflection acquires an additional phase shift by -6. This can be readily understood
by realizing that the Andreev reflection of an incident electron creates an additional
247
Cooper pair with phase 6. The phase shift -0 of the reflected hole simply compensates
for the phase of the new Cooper pair.
Conversely, if a hole is reflected, the wave function of the resulting electron
acquires the phase +9, with a similar interpretation. What matters for the Andreev
bound states is of course the net round-trip phase shift. If the two superconducting
banks have the same phase, the phase shifts by ±0 at the two banks cancel, but if there
is a phase difference between the two banks, it will make a contribution
to the round-trip phase shift, with the following sign rule: If, in Figure 3b, the left-hand
bank is bank #1, the minus-sign applies, otherwise the plus sign.
In order to retain the round-trip condition (6) in the presence of the phase shift
contribution A0, the latter must be compensated for by an opposite change in the phase
shift contribution associated with the ballistic flight through the semiconductor itself.
But this leads to a change of the energy of the Andreev bound states: A positive con-
tribution to the round-trip phase shift requires a lowering of the ballistic phase contri-
bution, and hence a lowering of the bound-state energy, while a negative contribution
raises the latter. Because of the sign difference in (7), in the presence of a nonzero
phase difference 02 - #i> the energies of the bound states will depend on the direction
of current flow in each state, in such a way that the states with a current flow in the
direction proper for a Josephson supercurrent will have a lower energy and hence a
higher thermal occupation probability, than those with a current flow in the opposite
direction. Hence, in this case there will be a thermodynamically stable net current flow,
even in the presence of scattering events.
Recall finally that a time-independent phase difference corresponds to zero bias
voltage. Hence the stable current is a true zero-resistance supercurrent, with a certain
maximum value, the critical current, for some particular value of the phase difference
e -0
2 l-
When the current through the device exceeds the critical current, a bias voltage
develops across the semiconductor, leading to the bending-over of the I-V characteristic
seen in Figure 2. This dissipative regime contains itself a rich variety of physical
phenomena, the discussion of which would again go beyond the scope of this paper; the
interested reader is referred to the literature, probably starting with a few existing
elementary review papers [4-6], which contain extensive references to key original
papers, including specifically to papers on the detailed theory for the various
phenomena.
248
♦ 200 -H-
1 /
V' V
Figure 4. Overall layout (bottom) of Nb grating structure, along with (top) a schematic
cross-section through a pair of Nb lines separated by a narrow stripe of InAs-AlSb
quantum well. All dimensions are in \lm.
1. Chang, L. D., Tseng, M. Z., Fork, D. K., Young, K. H., and Hu, E. L. (1992) Epitaxial MgO buffer
layer for YBa?Cu O thin films on GaAs, Appl. Phys. Lett. 60, 1753-1755.
2. Tseng, M. Z, Jiang, W. N., and Hu, E. L. (1994) Measurements and analysis of Hall effect of a two
dimensional electron gas in the close proximity of a superconducting YBa„Cu.O_ film, J. Appl.
Phys. 76, 3562-3565.
3. Likharev, K. K. and Semenov, V. K. (1991) RSFQ Logic/Memory Family: A new Josephson-Junction
Technology for Sub-Terahertz-Clock-Frequency Digital System, IEEE Trans. Appl Supercond. 1, 3-
28.
4. Kroemer, H. and Hu, E. (1996) "Semiconducting and Superconducting Physics and Devices in the
InAs/AlSb Materials System," in Nanotechnology, G. Timp, Ed., New York, AIP Press. In the press.
5. Kroemer, H., Nguyen, C, and Hu, E. L. (1994) Electronic Interactions at Superconductor-
Semiconductor Interfaces, Solid-State Electron. 37, 1021-1025. (Proc. MSS-6, Garmisch-
Partenkirchen, Germany, Aug. 1993).
6. Kroemer, H., Nguyen, C, Hu, E. L., Yuh, E. L., Thomas, M., and Wong, K. C. (1994) Quasiparticle
transport and induced superconductivity in InAs-AlSb quantum wells with Nb electrodes, Physica B
203, 298-306. (Proc. NATO Advanced Research Workshop on Mesoscopic Superconductivity,
Karlsruhe, 1994).
7. Feynman, R. P., Leighton, R. B., and Sands, M. (1965) The Feynman Lectures on Physics; Vol. 3:
Quantum Mechanics, Addison-Wesley, Reading. See Sec. 21-9.
8. Kittel, C. (1986) Introduction to Solid State Physics, Wiley, New York.
9. Tinkham, M. (1975) Introduction to Superconductivity, McGraw-Hill, New York.
10. de Gennes, P. G. (1966) Superconductivity of Metals and Alloys, Benjamin, New York.
11. Likharev, K. K. (1979) Superconducting weak links, Revs. Mod. Phys. 51, 101-158.
12. Andreev, A. F. (1964) The thermal conductivity of the intermediate state in superconductors, Sov.
Phys. JETP 19, 1228-1231.
13. van Houten, H. and Beenakker, C. W. J. (1991) Andreev reflection and the Josephson effect in a
quantum point contact, Physica B 175, 187-197.
14. Mead, C. A. and Spitzer, W. G. (1964) Fermi Level Position at Metal-Semiconductor Interfaces, Phys.
Rev. 134,713-716.
15. Nakagawa, A., Kroemer, H., and English, J. H. (1989) Electrical properties and band offsets of
InAs/AlSb n-Nisotype heterojunctions grown on GaAs, Appl. Phys. Lett. 54, 1893-1895.
16. Silver, A. H., Chase, A. B., McColl, M., and Millea, M. F. (1978) Superconductor-Semiconductor
Device Research, Future Trends in Superconductive Electronics, Charlottesville, VA, J. B. S. Deaver,
C. M. Falco, H. H. Harris, and S. A. Wolf, Eds., Am. Inst. Phys. Conf. Ser., vol. 44, Am. Inst. Physics,
pp. 364-379.
17. Clark, T. D., Prance, R. J., and Grassie, A. D. C. (1980) Feasibility of hybrid Josephson field effect
transistors, J. Appl. Phys. 51, 2736-2743.
18. Takayanagi, H., Akazaki, T, Nitta, J., and Enoki, T. (1995) Superconducting Three-Terminal Devices
Using an InAs-Based Two-Dimensional Electron Gas, Jpn. J. Appl. Phys. 34, 1391-1395.
19. Akazaki, T, Nitta, J., and Takayanagi, H. (1995) Superconducting Junctions using a 2DEG in a
Strained InAs Quantum Well Inserted into an InAlAs/InGaAs MD Structure, IEEE Trans. Applied
Supercond. 5, 2887-2891.
FIELD EFFECT TRANSISTOR AS ELECTRONIC FLUTE
Abstract.
When electron-electron collisions are more frequent than electron collisions with
impurities and phonons, electrons are described by hydrodynamic equations. Many new
interesting physical phenomena, such as wave instability, shock waves, turbulence, and
choking, should occur in this electron fluid. Plasma effects in is a High Electron
Mobility Transistor should allow us to design a new family of solid state devices - a
FET emitting far infrared radiation, an electronic flute, and a terahertz detector and
mixer. These devices should be able to push a three terminal device operation into a
much higher frequency range than has been possible for conventional, transit time
limited regimes of operation.
1. Introduction.
The development of silicon technology has led to a dramatic reduction in device sizes
(from 10 urn or so in the nineteen sixties to sub-0.1 urn in short silicon MOSFETs in
the late nineteen nineties). This reduction resulted in a decreasing number of electrons
in the device channel. However, the electron density has actually increased because
the gate voltage swing does not scale proportionally to the decreasing thickness of the
gate oxide. This is illustrated by Fig. 1, which shows the calculated dependence of the
surface and volume electron densities in the MOSFET channel on the gate length.
E
Ü
251
S. Luryi et al. (eds.), Future Trends in Microelectronics, 251-261.
© 1996 Khmer Academic Publishers. Printed in the Netherlands.
252
As seen from the figure, in a 0.1 (0.m Si MOSFET, the volume carrier concentration is
on the order of 1020 cm"3, which corresponds to a weakly degenerate and highly non-
ideal electron gas at room temperature. At room temperature, the electron thermal
velocity in silicon, vth = (3kBT/m) ~ l.lxlO7 cm/s, is of the same order as the Fermi
velocity. (In this estimate, we used the effective mass of density of states in Si, m ~
1.1 me, where meis the free electron mass.). The average distance between electrons is
R = n Aa = 40 A (of the same order as the Bohr radius), and the electron-electron
collision time is on the order of xee = Rs/ vth = 0.4xl0"14 s, much smaller than the
collision time with impurities and phonons (i. e. the momentum relaxation time, x =
\inmle ~ 0.2 ps at room temperature). Here e is the electronic charge and \in is the low
field mobility. Under such conditions, electrons in the MOSFET channel should
behave as a two dimensional (2D) electron fluid, i. e. they should be governed by
hydrodynamic equations.
The gate electrode in a FET (see Fig. 2) is separated from the channel by the
gate insulator [which is a silicon dioxide layer in a MOSFET and a doped or undoped
wide band gap semiconductor, such as AlGaAs in a typical High Electron Mobility
Transistor (HEMT)]. The surface concentration, ns in the FET channel is given by
ns=CU/e (1)
where C is the gate capacitance per unit area. Eq. (1) represents the usual gradual
channel approximation 2 which is valid when the characteristic scale of the potential
variation in the channel is much greater than the gate-to-channel separation, d.
2D U
9.
electron Gate Uate
InsL
a°°
^ <+-
+
U
Drain
Source
Fig. 2. Schematic structure of a FET. ' In a Ballistic FET, the gate length, L, is much smaller than the mean
free path, X, but much longer than the mean free path for electron-electron collisions, Xee.
We notice that eqs. (2) and (3) coincide with the hydrodynamic equations for shallow
water (see, for example, 3). Hence, the 2D electron fluid in a Ballistic FET should
behave like shallow water. In this hydrodynamic analogy, v corresponds to the fluid
velocity, and eUlm corresponds to gh where h is the shallow water level and g is the
free fall acceleration.
This analogy has profound consequences for understanding of interesting
physics of 2D electrons in the Ballistic FET. Phenomena similar to wave and soliton
propagation, hydraulic jump, and the "choking" effect 4, 5 should take place in this
hydrodynamic electron fluid. The effects of collisions, surface scattering, changes in the
channel cross section, and others may be also understood using this analogy.
The waves propagating in this 2D electron fluid are plasma waves, and their
dispersion law (which can be easily obtained from linearized equations (1) and (2)) is
similar to that for shallow water waves (or sound waves):
2. Electronic Flute.
Since, as discussed above, the behavior of the plasma waves in 2D systems is governed
by the same equations as for sound waves, resonant structures, similar to those in
musical instruments, may be realized for the plasma waves, and these waves can be
excited by a direct current just like wind musical instruments are excited by air jets.
However, the plasma wave velocity is much higher than the sound velocity and the
FETs are very small. As a consequence, the plasma wave frequencies are in the terahertz
range. These plasma waves are accompanied by a variation of a dipole moment created
by charges in the FET channel and mirror image charges in the gate and, hence, should
cause the emission of far infrared (terahertz range) electromagnetic radiation.
Plasma waves are similar to shallow water waves or to sound waves since
254
they have a linear dispersion law. In turn, shallow water behavior is similar to the
dynamics of a gas with pressure proportional to the square of the density, (see, for
example, 3). Thus, the nonlinear hydrodynamic equations for the 2D electron fluid are
similar to (but not identical with) the equations for a real gas, such as air. However, the
linearized equations describing small-amplitude plasma waves in a FET and sound waves
in a gas are identical. Since the linearized equations determine the instability threshold
for a steady flow (i. e. the wave generation threshold), the instability conditions for a
real gas and for a 2D electron fluid should be similar provided that the Reynolds
numbers and quality factors of resonance cavities are the same. Below, we will show
that these parameters for a HEMT structure may be of the same order of magnitude as
for a conventional flute.
Fig. 3a (from Ref. ,0) shows the schematics of a jet driven wind musical
instrument of a flute family. An air jet excites a resonant cavity formed by the pipe
closed at both ends. A similar structure can be realized using a modulation doped gated
device (see Fig. 3b). Hence, we call the part of this device shown below the dashed line
in Fig. 3b an electronic flute. The remaining gated portion of the device connected to
the drain is similar to the outside air space for a jet driven musical instrument. In the
electronic flute, a direct current flow excites plasma waves in the resonance cavity in the
same way as an air jet excites sound waves in an acoustic cavity.
Drain
Air jet
JL
Source
m
RSSonanc
(a) (b)
Fig. 3 a. Simplified diagram of a jet driven pipe musical instrument of a flute family. Arrows show the
direction of air flow.
b. Electronic flute. White areas show the gated region of the device with the 2D electron fluid in the
channel. The part of this device shown below the dashed line is the electronic flute. The top gated portion
of the device connected to the drain is similar to the outside air space for a jet driven musical instrument.
Arrows show the streamlines of the electron current, which flows between the source and the drain.
Unfortunately, the plasma waves are confined within the electron fluid and
cannot be directly enjoyed by a grateful audience. However, the plasma oscillations in
the cavity are coupled to electromagnetic radiation. Indeed, the plasma oscillations lead
to a variation of the dipole moment created by the electron charge in the channel and the
positive mirror charge in the gate metal. Since the device dimensions are small
compared to the wavelength, Kem, of an electromagnetic wave corresponding to the
plasma oscillation frequency (Xem ~ 100 |im), this dipole behaves like a point dipole
emitting electromagnetic radiation. The measurements of this radiation should provide
the means of detecting the excited plasma waves. A small size of the dipole leads to a
relatively weak coupling. However, this coupling can be greatly improved by using
special antenna structures or, better still, phase locked device arrays as discussed below.
255
Let us now compare the relevant parameters, which are the Reynolds numbers
and quality factors for a conventional flute and for an electronic flute (see Table 1). For
a conventional wind instrument, a critical dimension is on the order of 1 cm (which is
either the pipe diameter or the size of the embouchure hole). The characteristic
dimension for the electronic flute is on the order of 1 micron. We choose a relatively
small flow velocity of 10 cm/s for the conventional case. The upper bound for the
electron flow velocity is limited by the electron saturation velocity (below than
approximately 107 cm/s). The sound velocity in air is about 3xl04 cm/s while the
typical value of the plasma wave velocity in a FET is on the order of 108 cm/s.
Source
1, 2, 3, 4
resonance cavities
Electron Flow
Fig. 4. Array of electronic flutes with a more efficient coupling of plasma waves to electromagnetic
radiation. The plasma wave oscillations are excited in the peripheral resonance cavities by 2D electrons
o
flowing from the source to the drain.
Thus, a complete similarity between the plasma waves in a FET and sound waves led us
to believe in the possibility of realizing an electronic flute based on the excitation of
plasma waves by a direct current in gated modulation doped structures. This electronic
flute should operate in a terahertz range of frequencies and emit far infrared radiation.
U
d
Fig. 5. Schematic geometry of FET operating in detector mode.
Fig. 6 shows the fundamental resonant frequency versus the gate voltage swing.
Fig. 7 shows the calculated responsivity as a function of frequency.
10'
N 5 -0.1 pm
h-
> :
In small FETs with a high concentration of electrons in the channel, the electron
behavior is governed by hydrodynamic equations. Devices using plasma waves
propagating in this electron fluid, such as an oscillator, an "electronic flute", a detector,
and a mixer, should operate at much higher frequencies than those possible for
conventional, transit-time limited regimes of operation.
6. Acknowledgment.
The authors are grateful to Professor Robert Weikle for useful discussions and
comments. The work at the Ioffe Institute has been partially supported by the Russian
Government, by the International Science Foundation, and by the US Army through its
European Research Office. The work at the University of Virginia has been partially
supported by the US Army Research Office (Project Monitor Dr. John Zavada) and by
the Office of the Naval Research (Project Monitor Dr. Yoon Soo Park).
7. References.
1. Dyakonov M. I. and Shur M. S. (1993) Shallow Water Analogy for a Ballistic Field Effect Transistor.
New Mechanism of Plasma Wave Generation by DC Current, Phys. Rev. Lett, 71, 2465
2. Shur M. S. (1990) Physics of Semiconductor Devices, Prentice Hall, New Jersey
3. Landau L. D. and Lifshitz E. M. (1966) Fluid Mechanics, Pergamon, New York
4. Streeter V. L. and Wylie E. B. (1985) Fluid Mechanics, ch. 7, McGraw Hill, New York
5. Dyakonov M. I. and Shur M. S. (1995) Choking of Electron Flow - A Mechanism of Current Saturation
in Field Effect Transistors, Phys. Rev. B51, 14341
6. Allen, Jr., S. J., Tsui D. C, and Logan R. A. (1977) Phys. Rev. Lett., 38, 980
7. Tsui D. C, Gornik E., and Logan R. A. (1980) Solid State Comm., 35, 875
8. Dyakonov M. I. and Shur M. S. (1995) Two Dimensional Electronic Flute, Appl. Phys. Lett., August 21,
9. Dyakonov M. I. and Shur M. S. (1995) Detection and Mixing of Terahertz Radiation by Two
Dimensional Electronic Fluid, in the Proceedings of22d International Symposium on GaAs and Related
Compounds, Cheju, Korea, Aug. 28- Sep. 2
10. Fletcher N. H. and Rossing T. D. (1991) The Physics of Musical Instruments, Springer-Verlag, New
York
11. Shur M. S. and Eastman L. F. (1979) IEEE Trans. Electron Devices, ED-26,1677
12. Shur M. S. (1987) GaAs Devices and Circuits, Plenum, New York
13. Heiblum, M., Nathan M. I., Thomas D. C, and Knoedler C. M. (1985) Phys. Rev. Lett., 55,2200
14. Chao P. C, Shur M. S., Tiberio R. C, Duh K. H. G., Smith P. M., Ballingall J. M., Ho P., and Jabra A. A.
(1989) IEEE Trans. Electron Devices, ED-36,461
HETERODIMENSIONAL TECHNOLOGY FOR ULTRA LOW
POWER ELECTRONICS
M. S. SHUR, W. C. B. PEATMAN*, M. HURT, R. TSAI, T.
YTTERDAL, and H. PARK
University of Virginia, Charlottesville, VA 22903-2442, USA
^Advanced Device Technologies, Inc. .Charlottesville, VA 22903
Abstract.
1. Introduction.
All semiconductor devices utilize interfaces between different regions — ohmic, p-n
junctions, Schottky barrier junctions, heterointerfaces, interfaces between a
semiconductor and an insulator. Typically, these interfaces are planes separating
different regions. However, recently a new generation of semiconductor devices has
emerged. These devices utilize interfaces between semiconductor regions of different
dimensions and are called heterodimensional devices. An example of such an interface is
a Schottky barrier between a three dimensional (3D) metal and a two dimensional
electron gas. Other possible configurations include the interface between a two
dimensional electron gas and a two dimensional Schottky metal, an interface between a
one dimensional electron gas and a two dimensional Schottky metal, and an interface
between a one dimensional electron gas and a three dimensional Schottky metal.
Different heterodimensional Schottky contacts have several features in common
- smaller capacitance because of a smaller effective cross section and a wider depletion
region, a high carrier mobility related to properties of the two dimensional (2D) electron
gas, a smaller electric field, and a higher breakdown voltage. (A wider depletion region
is caused by fringing electric field streamlines.) These features make these devices very
promising for applications in ultra-high frequency varactors and mixers.
In this paper, we describe a new generation of devices utilizing Schottky
contacts between a metal and a 2Dimensional electron gas (2DEG). This new high
speed heterodimensional contact has unique characteristics which are particularly
promising for applications in the fields of millimeter wave electronics and high speed,
ultra low power integrated circuits. These devices include a new two terminal
263
S. Luryi et al. (eds.), Future Trends in Microelectronics, 263-268.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
264
heterodimensional Schottky diode (see Fig. 1 a) and three new heterodimensional
transistors which utilize a side-gate formed by plating gate metal into a trench etched
through the plane of the 2DEG.
Schottky
Schottky Contact
Contact Pad
4 6 8 10 12
Voltage (V)
Fig. 1. Schematic structure (a) and measured C-V characteristics (b) of a heterodimensional Schottky
diode. 1
Our heterodimensional transistors have a very small size (i. e. 0.5x0.8 \lrri2).
These small sizes, small gate-to-channel capacitances, and low parasitic capacitances
result in a nearly ideal performance with a very small number of electrons in the channel
(see Fig. 2).
r 100,000 CMOS
o 10,000
o
CD
CD 1,000
H—
1_ 100
CD
.a
E 10
z 1
0.1 0.2 0.3 0.4 0.5
Channel length (micron)
Fig. 2. Number of electrons, N, versus channel length for gate voltage swing, V„t = 0.5 V. For CMOS, N =
CjWLVg/q. For Heterodimensional Transistors, N = C^LVJq, where C,- = tgtjld is the gate capacitance
per unit area. C„ is the gate capacitance per unit length, W, L, and d are the channel width and length and
the gate dielectric thickness, respectively, e0e(- is the silicon dioxide permittivity, and q is the electronic
charge. In this calculation, we assumed d = 1/20, W = 10 L, C^ - ■ 10"10 F/m.
265
2. Heterodimensional Transistors.
Source
Gatel
Q Drain YZ
(a) (b)
Fig. 3. Schematic structure of a Resonant Tunneling Transistor (RTT) (a) and RTT current-voltage
characteristics 3 (b).
Side Side
Top gate
Drain gate gate
Gate 2 Gatel
2D gas
Source
Heterostructure
Buffer
(a) (b)
Fig. 4. Schematic Diagram of 2D MESFET (a) and coaxial MESFET (b) 3,4
266
In the coaxial MESFET, the electron gas is constricted from four sides and controlled
from three sides - by the top Schottky gate and by the two side Schottky gates. This
should allow us to achieve a very precise control of the electron density from a 2D gas
to a ID gas to a level of a few electrons in the channel. These devices are particularly
promising for low power, high speed integrated circuit applications because it
eliminates a narrow channel effect. 3 They also may achieve a high speed because of
reduced parasitic capacitance. In a conventional FET, the gate capacitance can be reduced
by either decreasing the device area or by increasing the gate-to-channel separation.
Both changes in the device dimensions lead to a substantial increase of CpIC since the
parasitic fringing capacitance decreases roughly in proportion to the gate periphery and
C decreases roughly in proportion to the gate area. Once the power delay product
becomes limited by CpAV2, where AV is the voltage swing, a further decrease in C only
leads to the deterioration in speed without any benefit for reducing the power
consumption. Hence, we conclude that the decrease in both parasitic and gate
capacitance is needed to achieve a low power technology. We also expect a coaxial
MESFET to have a very sharp pinch-off and an extremely small leakage current.
Fig. 5 compares qualitative distributions of the electric field streamlines in a
conventional HFET, in a 2D-MESFET, and in a coaxial MESFET. As can be seen
from this figure, most of the streamlines for a 2D MESFET and even more so for a
coaxial MESFET terminate on the gate electrode. Hence, the parasitic capacitance of a
2D-MESFET is smaller than that of a conventional device. This parasitic capacitance is
even smaller for a coaxial MESFET. These distribution shows that the
heterodimensional devices greatly reduce the detrimental narrow channel effect and,
hence, allow us to use narrower and lower power devices, reducing the gate capacitance
without a commensurate increase in the relative importance of the parasitic capacitance.
Of course, this does not solve the problem of driving the interconnects. The circuit
layout for low power electronics must have short interconnects, except for a few long
interconnects which have to be driven by special drivers. If the number of such long
interconnects is not large, then the share of the drivers in the total power budget can be
small or, at least, manageable.
Fringing streamlines
Fig. 5. Device cross sections and electric field streamlines in conventional HFET (a), 2D-MESFET (b), and
coaxial MESFET (c). 5
267
We demonstrated a 1 |im wide AlGaAs/InGaAs 2D MESFET having a peak
drain current and transconductance of 210 mA/mm and 210 mS/mm and the
subthreshold slope of 75 mV/decade corresponding to an ideality factor of 1.3 3 (this
value is comparable to that in the state-of-the-art 10 |Xm wide HFET). From this we
estimated the cutoff frequency to be about 21 GHz which is comparable to that of the
best 1 |J,m long HFETs.
By eliminating the narrow channel effect and reducing parasitic capacitances,
this new technology enables scaling the gate width to submicron dimensions to achieve
a large reduction in the power consumption without loss of speed performance. The
Schottky side gates modulate the width of the 2DEG channel and the current between
the source and drain. The gates are formed by etching through the plane of the
conducting layer and by electroplating Pt/Au onto the walls using resist as the mask.
Further details of the fabrication are described in 3. We fabricated both depletion mode
and enhancement mode devices. For a 0.5 ^.m wide channel device, the threshold
voltage was 0.0 V while the knee (ON) voltage was about 0.2 V. 6 Such device should
operate at less than 1 V bias. Based on our charge control model of the 2D MESFET,
we estimated the power-delay product of 0.1 fJ which is an order of magnitude smaller
than state-of-the-art technologies. We also observed a nearly zero threshold voltage shift
with temperature and almost total absence of DIBL in these devices.6
< VgS=0.6V
3 8
c 6 Vgs=0.5V
CO
7
3 4
c Vgs = 0.4V ■
'co 2
Q
•T"ft"TT-TWM-tTT-JvM-H
0 0.2 0.4 0.6 0.8 1.0
Fig. 6. Measured and simulated current-voltage characteristics of a 0.5 |xm 2D MESFET. "
3. Quasi-three-dimensional modeling.
4. Conclusion.
Our experimental results and device and circuit simulations show that the
heterodimensional technology holds promise of ultra low power operation at a
relatively high speed. It should also allow us to study the behavior of single electrons
in a wide temperature range.
5. Acknowledgment.
6. References.
1. W. C. B. Peatman, T. W. Crowe, and M. Shur, A Novel Schottky/2DEG Diode for Millimeter and
Submillimeter Wave Multiplier Applications, IEEE Electron Device Letters, vol. 13, No. 1, pp. 11-13, Jan.
(1992)
2. W. C. B. Peatman, E. R. Brown, M. J. Rooks, P. Maki, W. J. Grimm, and M. Shur, Novel Resonant
Tunneling Transistor with High Transconductance at Room Temperature, IEEE Electron Device Letters,
vol. 15, No. 7, pp. 236-238, July (1994)
3. W. C. B. Peatman, H. Park, and M. Shur, IEEE Electron Dev. Lett., vol. 15, No. 7, pp. 245-247, July
(1994)
4. W. C. B. Peatman, B. Gelmont, W. L. Grimm, H. Park, M. Shur, E. R. Brown, and M. J. Rooks,
Heterodimensional Schottky-Gate Devices, in Proceedings of 2D International Semiconductor Device
Research Symposium, Charlottesville, VA, December, pp. 427-430 (1993)
5. M. Shur, W. C. B. Peatman, H. Park, W. Grimm, and M. Hurt, Novel Heterodimensional Diodes and
Transistors, Solid State Electronics, Sep. (1995)
6. W. C. B. Peatman, R. Tsai, T. Ytterdal, M. Hurt, H. Park, J. Gonzales, and M. S. Shur, Sub-half-micron
Width 2D MESFET , unpublished
7. M. Hurt, M. S. Shur, W. C. B. Peatman, and P. B. Rabkin, Quasi-three-dimensional Modeling of a Novel
2D MESFET, IEEE Trans. Electron Devices, accepted for publication
8. H. Takato, K. Sunouchi, N. Okabe, N. Nitukyama, K. Hieda, F. Horiguchi, and F. Masuoka, High
Performance CMOS Surrounding Gate Transistor (SGT) for Ultra High Density LSIs, IEDM, pp. 222-225
(1988)
9. D. Hisamoto, T. Kaga, Y. Kawamoto, and E. Takeda, A Fully Depleted Lean-Channel Transistor
(DELTA) - A Novel Vertical Ultrathin SOI MOSFET, IEEE Electron Dev. Lett., vol. 11, No. 1, pp. 36-38
(1990)
LATERAL CURRENT INJECTION LASERS -
A NEW ENABLING TECHNOLOGY FOR OEICS
1. Introduction
At present, almost all semiconductor lasers use designs based on the vertical injection
of current into the active region. While this scheme has proven successful for discrete
laser diodes, it has several intrinsic drawbacks. First of all, there is an incompatibility
between vertical injection and the preference for lateral integration. Just as in electronics,
the evolution from discrete components to integrated circuits necessitates a change from
vertical to lateral devices.
Another drawback concerns the present use of a highly conductive substrate associ-
ated with the vertical injection scheme. This makes device-to-device isolation difficult
and can result in high parasitic losses and delays. In addition, the injected current must
pass through the cladding layers which provide optical and electrical confinement. The
resistance inherent to these wide bandgap layers is detrimental to the lasing threshold
and lasing efficiency, particularly under high injection conditions. This necessitates a
compromise when choosing the cladding layer material and thickness in order to provide
adequate optical confinement while avoiding excessive resistance.
Overcoming the resistance becomes critical in the case of Vertical Cavity Surface
Emitting Lasers (VCSELs), where the cladding layers are replaced by high reflectivity
Distributed Bragg Reflectors (DBRs). Since vertical current injection precludes the use
of insulating materials in the DBRs, many alternating layers of semiconductors with rel-
atively small differences in refractive index are needed. This results in a high resistance
path for the injected current and much lower lasing efficiency than is theoretically possi-
ble. Efforts have been made to alleviate this problem in VCSELs by using elaborate etch-
ing and recessed contacts at the expense of further complicating the fabrication.
Distributed FeedBack (DFB) lasers also suffer, since the current injection path passes
through the corrugated grating region of high interface defect density.
Lateral Current Injection (LCI) lasers have the potential to overcome many of these
limitations by making use of the under-explored lateral degree of freedom. In an LCI
laser, the planar geometry is well suited for OEIC applications. In addition, current does
not pass through the optical cladding layers or the substrate, allowing both these regions
269
5. Luryi et al. (eds.), Future Trends in Microelectronics, 269-278.
© 1996 Khmer Academic Publishers. Printed in the Netherlands.
270
Doped Substrate
B.)
Serai-insulating Substrate
C.) ■^"IP"^
Wave guiding layer
tc be designed for optimal optical confinement and made of wide bandgap semi-insulat-
ing material for reduced parasitic loss and capacitance. In this configuration the electrical
and optical designs are de-coupled and the cladding layers can be optimized separately
from the current injecting regions.
Despite these potential advantages and a considerable amount of experimental effort
devoted to this type of laser [1-26], the progress of LCI laser development has been
rather slow. In a large part this is due to the predominance of trial-and-error experimental
attempts with very little backing from theoretical analysis. The best results from the most
recent trials are still inferior to that of state of the art vertical injection designs. Typically,
the threshold current is high while the efficiency starts out low and then decreases with
increasing current.
It is difficult to determine the origins of these shortcomings from experimental mea-
surements alone. Many interdependent factors are involved, some of them uniquely new.
Starting from fundamental laser physics the job is no easier. First principle consider-
ations may be the same as for vertical injection, but new aspects of the electro-optic
interaction are involved which require self-consistent analysis. This is what we have
done using our 2D self-consistent Finite Element Light Emitter Simulator (FELES) [27].
In this paper we will first present a brief summary of the history and current state of
271
2. Progress to Date
Much of our work on LCI lasers has involved the theoretical analysis of various
designs using FELES [30]. In this section, we will present the results of our investigation
of the underlying physics of two recently reported BH-LCI structures that illustrate the
unique internal operating mechanisms of LCI lasers.
One of the advantages of BH-LCI lasers is the good carrier and optical confinement
achievable in both transverse directions. Since the cladding layers outside the active
region are not electrically active and serve only to guide the wave, the compositions and
widths of these layers as well as that of the active region can be designed to produce a
single transverse mode that exhibits an almost perfectly circular far field pattern (Figure
3) without compromising the electrical characteristics of the device.
Our simulations also reveal a key design issue that is rooted in fundamental device
physics unique to LCI lasers. The lateral gain profile in the QWs can be rather asymmet-
ric (Figure 3). This problem, which can only be seen clearly through simulation, origi-
nates from the disparity in mobilities between electrons and holes. To maintain quasi
charge neutrality, the higher mobility electrons tend to be pulled over to meet the lower
273
region (-0.85V), one might not expect leakage to be a problem. Indeed, this is true for
injected currents up to a few mAs as attested to by the low threshold currents of both the
experiment and simulation. Under high injection, however, the voltage drop becomes
sufficient to bias these parallel diodes near their turn-on voltages. This activates these
parasitic conduction paths and results in an exponential increase in leakage current and a
corresponding decrease in efficiency.
When the leakage problem is minimized, our analysis indicates that BH-LCI lasers
should be comparable to vertical injection lasers in terms of threshold and slope effi-
ciency. Another consideration which naturally arises is the device speed. This is a com-
plex and challenging subject involving such issues as 3D to 2D capture processes and
ambipolar transport in 2D space. Our initial investigation into this question indicates that
LCI lasers have, at least, the advantage of low parasitic delay.
Since the substrate can be made semi-insulating, the effective area of the n and p con-
tacts is small, and the separation between the contacts is relatively large, the capacitance
of LCI lasers should be much lower than vertical injection designs. In confirmation of
this, FELES simulations predict a zero bias capacitance of 0.8pF for the structure
reported in [26], compared to the 0.5pF measured experimentally. While it is difficult to
experimentally measure the capacitance under large forward bias, it was easily deter-
mined from FELES simulations and found to be approximately 15pF with the device
biased at twice the threshold current.
In addition, perhaps contrary to one's intuition, both experiment and simulation
results exhibit a low differential resistance (Figure 6). This can be explained by the fact
that although the conduction path through the active region is narrow, it is composed of
high mobility, low bandgap material. In contrast, the conduction path in conventional
lasers is quite wide but passes through the large bandgap, low mobility cladding layers.
4. Future Directions
As the basic design problems associated with LCI lasers are solved, many new and
exciting research possibilities will open up. The ridge waveguide, freed from the current
conduction requirements of vertical injection lasers, can be adapted to a variety of tasks.
Technological compatibility with FETs, HBTs, photoreceivers, and waveguides multi-
plies the number of options.
A dielectric ridge waveguide, in addition to its role in the optical confinement of the
laser mode, can also act as "photonic wiring". The laser output could be coupled into
these dielectric-loaded waveguides and, when extended over the wafer, they would act as
optical interconnects. These interconnects could route the optical output of LI lasers to
photoreceivers, phototransistors, and cross-connection networks. But can a dielectric
ridge of relatively low index deposited on high index semiconductor layers provide an
effective waveguiding structure? Via a series of investigations, we have found that the
answer is yes provided that the proper structure and layer designs are in place first. In
fact, the waveguiding ability of a dielectric ridge is so wavelength sensitive that it can be
a valuable new addition to our "tool box" of wavelength and mode control techniques
[31].
276
The basic structure can be extended to DFB
create a VCSEL by replacing the ridge DBR Wiring
with a DBR. Insulating dielectric material
can now be used for the DBR to achieve
high reflectivity with fewer layers than in
vertical injection designs. Adding Bragg
gratings to the base of the ridge would
produce a DFB laser. This grating can be
fabricated by traditional methods or opti-
cally. For the former, the stringent require-
ment for grating interface quality is
removed since no current will flow
through the region. For the latter with a
photorefractive dielectric ridge, the inter-
ference pattern created by two high power
laser beams incident on the ridge would Figure 7. Illustration of the post-processing choices
for an LCI laser. Since the ridge need not
"burn-in" the Bragg Reflector. This
be electrically conductive, it can be made
method results in low absorption and of dielectric material deposited after the
recombination losses and a laser operating semiconductor growth is complete. The
wavelength that could be chosen after the dielectric ridge can form a DBR for a
main part of fabrication is complete. Car- VCSEL, and DFB grating, or photonic
wiring.
rying this idea a bit further, even the type
of laser (cleaved facet, DFB, DBR, VCSEL, etc.) could be chosen after the material
growth is complete by depositing dielectric material and then patterning the appropriate
reflector (Figure 7).
By adding a metal gate on the top or on the side of the ridge waveguide, both the
emission wavelength and the power of the laser can be controlled. The former opens up
the possibility of yield enhancement by tuning. Devices on the same wafer inevitably
vary in wavelength because of material and/or process variations. With the ability to tune
the gain spectrum by gate field via the Stark Effect, we may be able to tune devices fab-
ricated outside the desired frequency range instead of discarding them. The success of
this approach could translate into a significant cost reduction in production. Once inte-
grated with logic circuitry, the gate could also be used to compensate for the change in
wavelength with operating temperature. The output wavelength could be monitored by a
circuit that would control the gate voltage to lock the laser output to a specified wave-
length. Therefore lasers using the same design and fabricated on a single wafer could be
set to different wavelengths for Wavelength Division Multiplexed (WDM) communica-
tions.
The possibility of gate-controlled emission power is even more exciting, as it may
realize the long-sought capacitive modulation (gain switching) of lasers via a third termi-
nal [32]. Implementing a gate in the space between the ridge and one of the contact
regions effectively turns the structure into an integrated laser-FET module. This type of
structure offers the potential of increased modulation speed due to minimized series
resistance.
277
By having the current flow in the lateral dimension, we have opened up the vertical
degree of freedom to many additional design possibilities for creating novel devices. For
instance, one can easily realize that multi-section lasers can readily be created along the
longitudinal direction without compromising the LCI design. This opens the possibility
of having multiple electrodes so that the gain or loss of each section can be tuned indi-
vidually.
5. Summary
References
1. Namizaki, H., Kan, H., Ishii, M., and Ito, A., (1974) "Transverse-Junction-Stripe-Geometry DH Lasers
with Very Low Threshold Current", J. Appl. Phys., 45, pp. 2785-2786.
2. Namizaki, H., (1975), "Transverse-Junction-Stripe Lasers with a GaAs p-n Homojunction", IEEE J. Quan-
tum Elec, 11, pp. 427-431.
3. Nagano, M. and Kasahara, K., (1977), "Dynamic Properties of Transverse Junction Stripe Lasers", IEEE J.
Quantum Elec, 13, pp. 632-637.
4. Kumabe, H., Tanaka, X. Namizaki, H., Ishii, M., and Susaki, W., (1978), "High-Temperature Single Mode
CW Operation with a Junction-Up TJS Laser", Appl. Phys. Lett., 33, pp. 38-39.
5. Ury, I., Matgalit, S., Yust, M., and Yariv, A., (1979), "Monolithic Integration of an Injection Laser and a
Metal Semiconductor Field Effect Transistor", Appl. Phys. Lett., 34, pp. 430-431.
6. Nita, S., Namizaki, H., Takamiya, S., and Susaki, W., (1979), "Single-Mode Junction-Up TJS Lasers with
estimated Lifetime of 106 Hours", IEEE J. Quantum Elec, IS, pp. 1208-1209.
7. Yang, Y.J., Lo, Y.C., Lee, G.S., Hsieh, K.Y., and Kolbas, R.M., (1986), 'Transverse Junction Stripe Laser
with a Lateral Heterobarrier by Diffusion Enhanced Alloy Disordering", Appl. Phys. Lett., 49, pp. 835-837.
8. Isshiki, K., Kaneno, N., Kumabe, H., Namizaki, H., Ikeda, K., and Susaki, W., (1986), "Ten-Thousand-
Hour Operation of Crank Transverse-Junction-Stripe Lasers Grown by Metal-Organic Chemical Vapor
Deposition", /. Lightwave Tech., 4, pp. 1475-1481.
9. Suzuki, Y, Mukai, S., Yajima, H., and Sato, X, (1987), Transverse Junction Buried Heterostructure (TJ-
BH) AlGaAs Diode Laser", Electronics Lett., 23, pp. 384-386.
lO.Ohta, J., Kuroda, K., Mitsunaga, K., Kyuma, K., Hamanaka, K, and Nakayama, X, (1987), "Buried Trans-
verse-Junction Stripe Laser for Optoelectronic Integrated Circuits", /. Appl. Phys., 61, pp. 4933-4935.
278
ll.DeFreez, R.K., Puretz, J., Orloff, J., Elliot, R.A., Namba, H., Omura, E., and Namizaki, H., (1988), "Oper-
ating Characteristics and Elevated Temperature Lifetests of Focused Ion Beam Micromachined Transverse
Junction Stripe Lasers", Appl. Phys. Lett., 53, pp. 1153-1155.
12.Shimoyama, K., Katoh, M., Noguchi, M., Inoue, Y, Gotoh, H., Suzuki, Y„ and Satoh, T, (1988), "Trans-
verse Junction Buried Heterostructure (TJ-BH) Laser Diode Grown by MOCVD", J. Crystal Growth, 93,
pp. 803-808.
13.Shimoyama, K., Katoh, M., Suzuki, Y, Satoh, X, Inoue, Y, Nagao, S., and Gotoh, H., (1988), "CW Opera-
tion and Extremely Low Capacitance of TJ-BH MQW Laser Diodes Fabricated by Entire MOVPE", Jpn. J.
Appl. Phys., 27, pp. L2417-2419.
14.Furuya, A., Makiuchi, M., Wada, 0., and Fujii, X, (1988), "AlGaAs/GaAs Lateral Current Injection Multi-
quantum Well (LCI-MQW) Laser Using Impurity-Induced Disordering", IEEE J. Quantum Elec, 24, pp.
2448-2453.
15. Ahn, D. and Chuang, SL., (1988), "A Field-Effect Quantum-Well Laser with Lateral Current Injection", /.
Appl. Phys., 64, pp. 440-442.
16. Yasuhira, N., Suemune, I., Kan, Y, and Yamanishi, M., (1990), "Selectively Doped Double-Heterostructure
Lateral Current Injection Ridge Waveguide AlGaAs/GaAs Laser", Appl. Phys. Lett., 56, pp. 1391-1393.
17.Honda, Y, Suemune, I., Yasuhira, N, and Yamanishi, M., (1990), "A New Optoelectronic Device Based on
Modulation-Doped Heterostructure: Demonstration of Functions as Both Lateral Current Injection Laser
and Junction Field Effect Transistor", IEEE Photonics Tech. Lett., 2, pp. 881-883.
18.Sin, Y, Hsieh, K.Y., Lee, J.H., and Kolbas, R.M., (1991), "Surface and Bulk Leakage Currents in Trans-
verse Junction Stripe Lasers", J. Appl. Phys., 69, pp. 1081-1090.
19.Honda, Y, Suemune, I, Yasuhira, N, and Yamanishi, M., (1991), "Continuous-Wave Operation of aLateral
Current Injection Ridge Waveguide AlGaAs/GaAs Laser with a Selectively-Doped Heterostructure", Jpn.
J. Appl. Phys., 30, pp. 990-991.
20.ZOU, W.X., Law, K.K., Merz, J.L., Fu, R.J., and Hong, C.S., (1991), "Laterally Injected Low-Threshold
Lasers by Impurity-Induced Disordering", Appl. Phys. Lett., 59, pp. 3375-3377.
21.Schaus, C.F., Torres, A.J., Cheng, J., Sun, S., Hains, C, Malloy, K.J., Schaus, H.E., Armour, E.A., and
Zheng, K, (1991), "Transverse Junction Vertical-Cavity Surface-Emitting Laser", Appl. Phys. Lett., 58, pp.
1736-1738.
22.Beyler, CA., Hummel, S.G., Chen, Q., Osinski, J.S., and Dapkus, P.D., (1991), "Low Threshold Current
Lateral Injection Lasers on Semi-Insulating Substrates Fabricated Using Si Impurity-Induced Disordering",
Electronics Lett., 27.
23.Hihara, M., Hirata, T, Suehiro, M., Maeda, M., and Hosomatsu, H., (1991), "Fabrication of GaAs/AlGaAs
Lateral Current Injection Quantum Well Laser", Extended Abstracts of the 1991 Int. Conf. on Solid State
Dev. and Mat., Yokohama, pp. 735-736.
24.Evaldsson, P.A., Taylor, G.W, Cooke, P., Burrus, C.A., and Tell, B., (1992), "Small Signal and Continuous
Wave Operation of the Lateral Current Injection Heterostructure Field-Effect Laser", Appl. Phys. Lett., 60,
pp. 1697-1699.
25.Kawamura, Y, Noguchi, Y, and Iwamura, H., (1993), "Lateral Current Injection InGaAs/InAlAs MQW
Lasers Grown by GSMBE/LPE Hybrid Method", Electronics Lett., 29, pp. 102-104.
26. Oe, K., Noguchi, Y, and Caneau, C, (1994), "GalnAsP Lateral Current Injection Lasers on Semi-Insulating
Substrates", IEEE Photonics Tech. Lett., 6, pp. 479-481.
27.Tan, G.L, Lee, K., and Xu, J.M., (1993), "Finite Element Light Emitter Simulator (FELES): A New 2D
Software Design Tool for Laser Devices", Jpn. J. Appl. Phys., part 1, 32, pp. 583-589.
28.Laidig, W.D., Holonyak, N., Camras, M.D., Hess, K., Coleman, J.J., Dapkus, P.D., and Bardeen, J., (1981),
"Disorder on an AlAs-GaAs Superlattice by Impurity Diffusion", Appl. Phys. Lett., 38, pp. 776-778.
29.Tan, G.L., Xu, J.M., and Shur, M., (1993), "GaAs/AlGaAs Double-Heterostructure Lateral P-I-N Ridge
Waveguide Laser", Optical Engineering, 32, pp. 2042-2045.
30.Suda, D.A., Lu, H, Makino, X, and Xu, J.M., (1995), "An Investigation of Lateral Current Injection Laser
Internal Operation Mechanisms", to be published in IEEE Photonics Tech. Lett..
31.Pavlidis, D., Sweeny, M., Anis, H., and Xu, J.M., (1995), "Bragg Reflectors for Mode Control in Directions
Orthogonal to the Bragg's Periodicity", to be presented at the 1995 Canadian Semiconductor Conference.
32.Sun, C.C. and Xu, J.M., (1989), "Observation of Capacitive Modulation of Bipolar Current in Poly-Si
Gated Lateral PIN Structures", Appl. Phys. Lett., 54, pp. 1875-1877.
WIDE BAND GAP SEMICONDUCTORS. GOOD RESULTS AND
GREAT EXPECTATIONS.
M. S. SHUR
Department of Electrical Engineering
University of Virginia, Charlottesville, VA 22903-2442, USA
Abstract.
We will review properties of wide band gap semiconductors, which make them
superior materials for many electronic and optoelectronic applications. These semi-
conductors should allow us to achieve a very high on-to-off ratio in transistors,
implement nonvolatile solid state memories, and develop new optoelectronic and
optical devices for visible and ultraviolet ranges as well as electronic and optoelectronic
systems operating in a harsh environment and/or at elevated temperatures.
Technological difficulties, relatively low mobility values, and problems related to
contacts and traps make the realization of this great potential a challenge. We show
that many of these difficulties can be alleviated in AlGaN/GaN Heterostructure Field
Effect Transistors (HFETs), which use superior transport properties of the two
dimensional electron gas in wide band gap semiconductors. AlGaN/GaN HFET's,
which have been fabricated on a transparent sapphire substrate, are very sensitive to
ultraviolet light. Hence, they can be also used as solar blind ultraviolet photodetectors.
1. Introduction.
Since amorphous silicon has a relatively wide energy gap (the energy gaps are
1.12 eV for silicon, 1.42 eV for GaAs, and 1.71 eV for amorphous silicon), amorphous
silicon thin film transistors have the largest on-to-off ratio, in spite of a very low on-
current. Wide band gap semiconductors (such as GaN with the energy gap of 3.4 eV)
should have very low leakage currents [since the intrinsic carrier concentration, n,-, is
proportional to exp(-Eg/2kBT)]. They have extremely large breakdown voltages (since
in order to cause an impact ionization event, an electron or a hole has to obtain an
energy in excess of Eg from the electric field). These materials are thermally and
chemically stable and uniquely suited for applications in electronic circuits and systems
operating in a harsh environment and/or at high temperatures.
o <D O) A0 On-state
•*-* -o Source Drain
o III
_}
TJ -o
r c Distance
o CO
O QJ
c (1) Off-state
o O)
•*-» TJ
o III
_J
T3 TJ Source
r C
n CO
o en
Distance
Fig. 1. Band diagram for a Field Effect Transistor.
JR=(1, (l)
*pNd
where q is the electronic charge, «,- is the intrinsic carrier density, Nd is the doping
density in the «-type region, Dp is the hole diffusion coefficient, xe is the effective
lifetime, and
(2)
i qNd
is the thickness of the depletion region. Here V^ is the built-in voltage and e^ is the
dielectric permittivity of the semiconductor.
In wide band gap materials, the second term in the right-hand part of eq. (1) is
dominant. In cubic SiC at room temperature, the theoretical value of «,• is as low as
10-6 cm-3! For an estimate, let us assume certain values for mn and mp, independent of
the energy gap, since this assumption will not change the order of magnitude of the
resulting concentration. Let us take mn = 03me and mp = 0.6me, to be specific. Then
we find
1012
Si\
CO
108
£ GaAs
o
104
10° i 1
0.5 1 1.5 2 2.5
Energy gap (eV)
Fig. 3. Intrinsic carrier concentration versus energy gap at room temperature. (The values of ns for Si and
GaAs are approximate since we used the same values of mn and m„ in the calculation of this
dependence.)
282
Choosing the device volume to be 1 X 0.1 X10 |J.m3 and assuming a generation time of
1 ns, we obtain
( E }
IunkaM
1
leakage^ = 2.14 x 1CT7 x T3/2(K) x exp (4)
2kBT
The resulting dependence of the minimum leakage current on the energy gap for room
temperature is shown in Fig. 4. However, experimentally, such low values of the
saturation current density have not yet been observed, and the conventional theory of a
p-n junction is not even valid when n[ is very small. Nevertheless, these estimates
clearly illustrate the potential of wide band gap semiconductors for applications in
nonvolatile memories and integrated circuits operating both at room temperature and at
elevated temperatures.
< Si \
r io-14 y GaAs
c
CD
\
t 10-18 \
i io-22
CO
V- 10 -26 SiC*
CD
C
CD 0 1 ;2 3
CD
Energy gap (eV)
Fig. 4. Generation current (per 1 u,m3 of the depletion region volume) versus energy gap
(at room temperature).
Wavelength (\im)
2 1 0.8 0.5 0.4 0.3 0.2
A1N
J l_l III I
1
ZnS
GaN
SiC (2H)
CO
SiC(6H) c
ZnSe o
ZnTe D.
SiC (3c) CO
GaP CD
AlAs "1 OH
AIP CD
CdTe 1
AlSb
GaAs CD
InP- >
Si -
Ge~ CO
GaSb
InAs ~ "CD
InSb -
T I I T i i i i—i i i i
0 2 3 4 5
Energy Gap (eV)
Fig. 5. Energy gaps of semiconductor materials compared with spectral sensitivity of human eye.
CM Wavelength (pm)
E 1 0.75 0.5 0.25
o 1 1 1
CO
0.15
X5
CO
CD -
0.1 GaN/^^
CD
»+—
o Si£.
0 Oh
CD
E
Si L. .1 ■ ■■I
13 2 3 4 5
Energy gap (eV)
Fig. 6. The number of bits per cm^ as a function of the photon energy and the corresponding wavelength,
>-\
X, (assuming the minimum pixel area of 10 Ar).
284
In order to take advantage of the properties of wide band gap semiconductors, one has to
develop the material, device, and integrated circuit technologies. So far, this ambitious
goal has been fully achieved only for one semiconducting material - silicon - and
partially achieved for gallium arsenide. Since Si has a huge cost and integration scale
advantage over GaAs, the jury is still out on whether GaAs will be able to win a fairly
large market share in high speed and low power applications or will remain an
important niche technology. Since GaAs has been studied intensively since the
nineteen fifties, we can expect that the realization of the full potential of wide band gap
semiconductors will take at least a few years if not a few decades.
At least four families of wide band gap semiconductors have attracted interest
of researchers worldwide - diamond, II-VI semiconductors such as ZnS or ZnSe, SiC
and related materials, and GaN and related materials.
Silicon carbide is one of the first semiconductor materials to be discovered. As
early as 1907, Round observed electroluminescence in silicon carbide, which he reported
in his article published in Electrical World, Vol. 19, p. 309, 1907. In 1955, Lely
developed a new technique of SiC growth, and the studies of silicon carbide started to
develop on a more sound basis. However, only recently, have practical applications of
silicon carbide devices become a reality with the development of modern epitaxial
techniques, primarily by groups in Russia (LETI and A. F. Ioffe Institutes in Sankt
Petersburg), US (NCSU and CREE research), and in Germany (Erlangen).
Silicon carbide exists in more than 170 different polytypes. The properties of
the polytypes are so different that, in fact, SiC may be more accurately considered as a
group of closely related materials. Depending on the polytype crystal structure, the
energy gap of silicon carbide varies from 2.2 to 3.3 eV. It has a predicted electron
saturation drift velocity, vs, of 2X107 cm/s (approximately two times larger than in
silicon), a breakdown field larger than 2,500 to 5,000 kV/cm (compared to 300 kV/cm
for silicon), and a high thermal conductivity of 3.5 W/cm°C (compared to 1.3 W/cm°C
for silicon and 0.5 W/cm°C for GaAs). These properties make SiC important for
potential applications in high-power, high-frequency devices as well as in devices
operating at high temperatures and/or in a harsh environment. Applications of SiC
include high-power devices, microwave devices (both avalanche diodes and microwave
field effect transistors), and optoelectronic devices such as light-emitting diodes
covering the visible electromagnetic spectrum and even the ultraviolet range.
Recently, new solid-state solutions of AIN/SiC/InN/GaN have been
demonstrated. This exciting development opens up the possibility of a new generation
of heterostructure devices based on SiC. Solid-state solutions of AIN-SiC are also
expected to lead to direct gap ternary materials for UV and deep blue optoelectronics,
including the development of visible light-emitting diodes and lasers.
All basic device elements - from ohmic contacts to Schottky diodes and p-n
junctions - and most of the semiconductor devices - from field effect transistors to
bipolar junction transistors, thyristors, and light-emitting Diodes - have been
demonstrated in materials ranging from different polytypes of SiC 2 to the GaN/AIN
material system. 3
In this lecture, we will limit ourselves to GaN, since this material and related
semiconductor can excel in both electronic and optoelectronic applications.
285
The Monte Carlo calculations show that the peak field in GaN is very high
(on the order of 100 kV/cm compared to approximately 3.5 kV/cm in GaAs); see Fig.
7. The reason for such a shape of the velocity-field characteristic is a large intervalley
separation and a very large energy of polar optical phonons in GaN (nearly three times
higher than in GaAs). In high electric fields, the electron velocity in GaN is only
weakly dependent on temperature (see Fig. 7 b).
— 3
03
E GaAs/^\GaN
o
N
o
2 - 1/ SiC^V--
Is
>—
Velocit
i
fs\
o
Compared to GaAs, GaN has a relatively low mobility. The most important scattering
mechanisms in GaN are polar optical scattering, ionized impurity scattering, and
piezoelectric scattering. In principle, the ionized impurity scattering may be nearly
totally suppressed in the two dimensional electron gas (2DEG) in a modulation doped
heterostructure, since the donors are located in a barrier layer, away from the 2DEG. In
practice, this is very difficult to achieve in an AlGaN/GaN modulation doped structure.
However, the calculation of the electron mobility limited by combined optical polar,
piezoelectric, and acoustic scattering gives an upper bound for the 2DEG mobility. The
results of such a calculation of the electron drift mobility are shown in Fig. 8. * (The
Hall mobility is a higher than the drift mobility.)
Fig. 9 shows the dependencies of the electron drift mobility in GaN on the
electron carrier concentration for different compensation levels at 300 and 77 K. As can
be seen from the figure the electron mobility in 2DEG can be substantially enhanced
due to the screening of the piezoelectric scattering. This enhancement is a feature
specific for GaN, which is a very strong piezoelectric. We notice large calculated
values of the drift mobility for uncompensated material.
In our recent paper , we reported on the measurements of the Hall mobility in
bulk GaN and in the 2DEG at the GaN/AlGaN heterointerface. These data clearly
demonstrate the difference between the electron transport in 2DEG and bulk GaN. Fig.
286
10 shows the temperature dependence of the electron Hall mobility in GaN for the
2DEG and bulk GaN. Solid lines and open dots show the calculated dependence and our
experimental data, respectively. For the bulk calculation, we used n = 2xl016 cm"3, A^
= 2xl017 cm"3 and for the 2DEG calculation, n = 5xl017 cm"3, NT = 6.5xl016 cm"3.
> 105
CM
3 104
zoelectric
is io3;
o
Combined
2
10 _i i i i
Fig. 8. Mobility limited by polar optical, piezoelectric, and acoustic scattering versus temperature-
2000 r , 20000 r , T = 77 K
T = 300 K
>1500 15000
E
o
1000 10000
X3
O
500 5000
(a) (b)
Fig. 9. Electron drift mobility in GaN versus electron carrier concentration for different
compensation levels at 300 K (a) and 77 K (b). 5
287
CO
6000
CM
E 4000
o
2000
!5
o 0
t 100 300 500
Temperature (K)
7
Fig. 10. Measured and calculated Hall electron mobility in bulk GaN and in the 2DEG.
The results of the calculation are in good agreement with our experimental data
for the 2DEG. For the bulk GaN, the theory overestimates the electron mobility at
elevated temperatures where the polar optical polar scattering should be the dominant
scattering mechanism. As was pointed by Professor Mishra 7, the most likely reason
for this disagreement is the nonuniform distribution of the electron concentration and
compensation ratio in the GaN film, with a strongly compensated layer near the
channel/buffer interface. However, the qualitative agreement with experimental data is
good enough to illustrate the mobility enhancement in the 2DEG
25 °C 200 °C
GaAs MESFET
Fig. 13. Computed device transconductance in the saturation region versus gate bias for three devices: 0.5
Urn gate GaAs MESFET, 0.5 um gate AlGaAs/GaAs HFET, and 0.25 um AlGaN/GaN HFET. 4
289
300 rr
Unique optical and electronic properties of the GaN/AlGaN material system open
up numerous opportunities for visible-blind optoelectronic devices. These devices have
a high sensitivity and a large gain-bandwidth product and can be integrated with
GaN/AlGaN field effect transistors which have already demonstrated an operation at
microwave frequencies. A transparent sapphire substrate makes AlGaN/GaN HFETs
well suited for optoelectronic applications.
The HFET photodetector is based on a 0.2 micron gate AlGaN/GaN HFET 9
and utilizes a shift in the threshold voltage caused by the light generated carriers (see
Fig. 15). These results show that unique optical and electronic properties of
GaN/AlGaN material system open up numerous opportunities for visible-blind
optoelectronic devices. These devices could have a high sensitivity and a large gain-
bandwidth product and can be integrated with GaN/AlGaN field effect transistors for
applications in optoelectronic integrated circuits.
Dark
— Illumination,
-3 -2 -1
Gate voltage (V)
9
Fig. 15. I-V characteristics of AlGaN/GaN HFETs in the dark and under light.
290
5. Conclusions.
Wide band gap semiconductor devices in general and GaN-based devices, in particular,
have already demonstrated an impressive performance. Their excellent transport and
optoelectronic properties should allow us to achieve excellent performance in harsh
conditions and/or at elevated temperatures. Large band discontinuities in these materials
should allow us to obtain very high densities of 2DEG (up to 2xl013 cm"2 for GaN).
2DEG should have much better transport properties than bulk electrons. This has
already been confirmed by a good performance of AlGaN/GaN HFETs.
6. Acknowledgment.
The work at the University of Virginia has been partially supported by the ONR
(Project Monitor Max Yoder).
7. References.
1. Shur, M. (1995) Introduction to Electronic Devices, John Wiley and Sons, New York
2. Keiner, G. and Shur M. (1995) SiC Devices, in "Properties of Silicon Carbide", G. Harris, Editor, M.
Faraday House, IEE, England
3 Strite, S. and Morkoc, H. (1992) J. Vac. Sei. Technol. BIO 1237
4 Shur, M. and Khan, M. A. Electronic and Optoelectronic AlGaN/GaN Heterostructure Field Effect
Transistors, unpublished
5 Shur, M., Gelmont, B., and Khan, M. A„ High Electron Mobility in Two Dimensional Electrons Gas in
AlGaN/GaN Heterostructures, unpublished
6 Khan, M. A., Chen, Q, Sun, C. J., Shur, M. S., and B. Gelmont, B, 2D-Electron Gas in GaN-AlGaN
Heterostructures Deposited Using TMAA as the Aluminum Source in Low Pressure Metalorganic
Chemical Vapor Deposition, unpublished
7 Mishra, U. (1995) Private communication, June 21
8 Khan, M. A., Shur, M. S., Kuznia, J. N, Burm J., and Schaff W. (1995) Temperature activated
conductance in GaN/AlGaN heterostructure field effect transistors operating at temperatures up to 300
°C, Applied Physics Letters, 66,1083
9. Khan, M. A., Shur, M. S., Chen, Q„ Kuznia, J. N., and Sun C. J. (1995) Electronics Letters, 31, 398
GaN and Related Compounds for Wide Bandgap Applications
Dimitris Pavlidis
1. Introduction
GaN and related compounds are wide bandgap semiconductor materials with
great potential for optoelectronic applications from blue to ultraviolet
wavelengths, and high-power, high-temperature devices. GaN can be
crystallized in either hexagonal (wurtzite) or cubic (zincblende) structure
depending on the substrate symmetry and growth conditions. In certain cases
both structures may co-exist because of the small difference in energy of
formation. High-quality wurtzitic GaN has been grown successfully on a
variety of substrates, in particular on the basal plane of sapphires. However,
cubic structures possess in principle superior electronic properties i.e. doping
efficiency and high-speed transport and allow easy cleaving, as necessary for
devices such as lasers [1],[2]. Although the traditional substrate used for nitride
material growth is sapphire, it is consequently desirable to explore the
possibility of using substrates such as silicon or GaAs. In addition to allowing
cubic material growth, this could lead to reduction of interfacial defects and
impurities as well as, integration of GaN with Si or GaAs-based devices.
The key binaries of nitrides are GaN, A1N and InN. Their bandgap energy
varies from 1.9eV for InN, to 6.2eV for A1N as shown in the diagram of Fig. 1
where the bandgap energy is plotted for various materials as a function of the
lattice constant. Shown in the same figure are other commonly used III-Vs such
as GaAs and InP. As one can see the lattice constant of nitrides is much smaller
than that of III-Vs or Silicon and none of the traditionally used semiconductor
substrates is compatible with them from the point of view of lattice matching.
The major difficulty encountered in the development of high quality nitride
films lies in fact on the lack of suitable substrates which leads to the need for
heteroepitaxial solutions.
>
a.
a
oi
>>
O)
L.
<D
C
UJ
-l—i—i—i—i—I—i—I—I—I—I—I—i—i—I—i—I—r-
4 5 6
Lattice Constant [A]
Figure 1. Bandgap energy vs. Lattice constants of nitrides and other commonly
used semiconductor materials
The most popular substrate used for nitride growth is sapphire (a=4.758Ä,
c=12.99Ä). It is stable up to the very high temperatures (950°C to 1050°C)
normally used for nitride growth but has a rather low thermal conductivity of
293
0.5W/cmK compared with other substrates such as 3C-SiC or 6H-SiC
(4.9W/cmK) and Si (1.5W/cmK). By way of comparison GaAs has similar
thermal conductivity as sapphire and the corresponding values for GaN, A1N
are 1.3W/cmK, 2.0W/cmK. These differences make sapphire not an optimum
choice for applications where heat needs to be dissipated very efficiently, as for
example high-power/high -temperature electronics. Differences also exist
between the values of coefficients of thermal expansion of the nitrides and the
substrates on which they are deposited, leading to difficulties upon cooling
down at the end of their growth.
The lack of suitable substrates for nitride growth prompted research on the
development of buffer technologies. GaN or A1N can be used for this purpose.
A study of the role of A1N buffer [6] showed that following the low
temperature growth of a 500Ä thick A1N layer, GaN nucleation sites are
generated with the same orientation as the substrate. This promotes the lateral
growth of GaN due to the decrease of interfacial free energy between the
substrate and the epitaxial GaN film, leading eventually to good quality bulk
GaN. The optimum buffer thickness has been reported to be small (-200Ä) in
case of GaN buffers. The use of a buffer layer helps in improving the crystal
quality as demonstrated by the reduction of Full-Width-Half-Maximum
(FWHM) values of X-Ray Diffraction (XRD) spectra of GaN grown with and
without buffer layers [7]. Moreover the introduction of a GaN or A1N buffer
leads to reduction of the residual carrier concentration and improvement of the
mobility of bulk GaN grown on top [8].
294
Various growth techniques have been employed for the growth of nitrides.
These include Molecular Beam Epitaxy (MBE), Metalorganic Chemical Vapor
Deposition (MOCVD) and Hydride Vapor Phase Epitaxy (HVPE). Electron
Cyclotron Resonance (ECR) used in conjunction with MBE allows activation
of molecular nitrogen N2+ and low temperature growth under ultra high
vacuum conditions leading to reduced autodoping [3]. MOCVD growth is in
most cases carried out at low pressure of -0.1 atm using NH3 gas and TMGa.
Growth temperatures for nitrides on sapphire substrates are in the order of
1050°C, while lower temperatures (~600°C) are sufficient for growth on GaAs.
A common feature of all nitride films is their relatively high background
concentration which has for long being associated with nitrogen vacancies
(VN). AS the growth temperature increases the background concentration is
normally found to decrease but VN increases [4]. Thus, VN alone cannot
account for the observed high background electron concentration and other
effects such as impurity incorporation i.e. O or Si may be present. HVPE has
been reported to permit growth of "bulk-like" GaN substrates opening therefore
the possibility of homoepitaxial GaN growth [5]. A sputtered ZnO buffer layer
has been employed in this case between the sapphire substrate and the GaN
and was removed at the end of growth leading to thick (~600um) GaN films.
Growth took place at high growth rates (80-130um/hr) and high crystalline
quality GaN was grown on top of these substrates.
6.00-1 — —1
<xh
1
lil
11
i
'
; I
!
2.00-ffln^B 35M
KU 1
i
3o.oouyHEi
20.00JVWM
lo.ootepJj
I
I 0
1K5 #"--,
to
30.00
Ö 0 0
0 b b
0 0
p-doping of nitrides has primarily been based on the use of Cp2Mg sources.
The Mg dopant introduced in this way was reported to require activation by
Low-Energy-Electron-Beam-Irradiation (LEEBI) and hole concentrations in the
order of 3xl0'18cm"3 have been demonstrated in this way [19]. Hydrogen
acceptor compensation was proposed as possible reason for the high resistivity
observed with Mg doping. LEEBI allows conversion of Mg-H complexes to the
expected Mg doping. Moreover Mg seems to be more sensitive to LEEBI than
other group II metals due to the lack of d-electrons and thus possibility of
resulting in shallow rather than deep acceptors. Carbon doping has also been
attempted more recently using CCU sources and hole concentrations of
3xl0'17cm-3 have been achieved [20]. n-doping is usually achieved by SiH4 or
Si2H6 and high electron concentrations of mid 10'19cm~3 have been
demonstrated with smooth, mirror like surfaces.
The large (~1.5eV) satellite valley separation from the central valley minimum
in GaN permits large peak velocities to be attained. Monte-Carlo simulations of
electron transport in GaN suggested peak velocity values as high as
2.7xlO'7cm/sec for GaN doping levels of ~10'17cnr3 and showed that
intervalley transitions in high electric fields play an important role in spite of
296
the large separation between the central and upper valleys [21]. This leads to
the possibility of realizing electronic devices with reasonably high-frequency
characteristics unlike what was initially expected for nitride semiconductors.
10
0 200 400 600 800 1000 0 200 400 600 800 1000
Electrical contacts on nitrides are still in their infancy but significant progress
has been made during the last few years in the understanding of contact
formation and their quality improvement. Al is commonly used for ohmic and
Au for Schottky contacts [22]. The Schottky barrier height achieved using Au is
0.8 to 0.9eV and ideality factors close to 1 are possible. Annealing at 575°C
has, however, been reported to lead to contact degradation as manifested by
change to ohmic behavior and is probably related to the presence of Au
diffusion [22] Specific contact resistivities of 8'10-6ßcm2 have been achieved
using Ti/Al [23]. The low contact resistance was speculated to be due to the
solid phase reaction between Ti and GaN forming TiN. If the nitrogen is then
297
out-diffused from the GaN, N vacancies and thus heavy doping would be
present at the surface allowing realization of tunneling contacts.
N-InGaN Layer
N-GaN Layer
A
GaN Buffer Layer
Sapphire Substrate
Very promising characteristics for blue LEDs have been presented by Nichia,
Japan [13]. A double heterostructure consisting of p-GaN/n-InGaN/n-GaN has
been used for this purpose. Its cross section together with the output power
characteristics of the diode are shown in Fig.4 [13]. Emission was detected at
430nm. A maximum output power of about 125uW at 20mA can be seen from
these characteristics which is higher than the 60jxW achieved by Zn(S,Se) at the
298
same current. The quantum efficiency of the GaN LED was also found to be
superior to that of ZnSe; 0.22% and 0.10% for GaN and Zn(S,Se) respectively.
li! 6nmAIN I
3 pm GaN
5 io - "'
40 nm AIN - A J_
Sapphire
:—J ' ' ' 1 1 1 1
2 3 4 S 6 7
Drain Voltage (V)
Figure 5. Cross-sectional view and DC high-temperature (300°C)
characteristics of AIN/GaN HEMT.
The results discussed in the preceding sections show that nitride research has
tremendously progressed during the last few years. Material quality has been
improved and promising device characteristics have been demonstrated for both
optical and electronic components. A major obstacle in nitride development has
been the lack of lattice matched substrates. Various possible solutions exist to
face this difficulty. The traditional approach used up to now is the development
of suitable buffer technology to allow good heteroepitaxial growth. Other
alternatives would be the development of growth techniques for bulk GaN
substrates and the study of "compliant" substrate technologies which would
confine dislocation within an initial "soft" buffer layer.
300
Growth techniques may be possible to take advantage of the availability of
sources such as "supersonic jets" in order to control the energy of injected
species and thus improve the quality of the layers by achieving a smaller
window of energy spread and lower growth temperatures. Improvements are
also still necessary in terms of residual doping control and intentional doping.
Basic studies are necessary in this area to fully understand the basic physical
properties of nitrides and their relation to the observed electrical and optical
characteristics.
Although encouraging results have been obtained for both optical and
electronic devices, significant improvements are still necessary before use of
such components can be made. This applies to LEDs, lasers, photodetectors
operating in the blue spectrum, as well as, MESFETs and HEMTs built on
nitrides for high-power/high-temperature operation. Material quality
improvement and controlled doping are parameters that impose difficulties in
such developments. Basic transport property studies and experiments on
heterostructure formation and control are essential for the success of such
developments.
New applications may emerge from the above discussed nitride developments.
For example, photoemission studies of A1N showed negative electron affinity
(NEA) characteristics [25]. These could be extremely interesting for the
development of electron emitters. A possible application of them could be
microwave tubes where cold cathodes could be realized based on the NEA
characteristics of nitrides.
Overall, the success of recent nitride developments during the last five years
sets up a solid base for future developments of optical devices in the blue
spectrum and electronic components for high-power/high-temperature
applications. The synergy between material, chemistry and electrical/electronic
engineering expertise is expected to lead to further progress, as necessary for
practical use of such components in various applications.
7. Acknowledgment
8. References
1. Introduction
Two main approaches exist for achieving a compact blue laser source. Using
the presently avälable high power infrared diode lasers (GaAs technology), nonlinear
conversion by second harmonic generation has provided tens of mW of coherent
blue emission, power levels required for optical writing (reading is performed at mW
level). Such equipment, which is becoming commercially available, is however quite
expensive due to the complex optoelectronic engineering designs which are also
excessively bulky e.g. for a CD-ROM. The second, direct approach is to utilize a wide
bandgap semiconductor such as ZnSe or GaN as the basis of an eletrical injection
laser - a task which until recently appeared rather hopeless.
On the other side of the seesaw, the wide gap semiconductors enjoy
fundamentally large interband optical cross sections - an obvious asset for blue light
emitters. Excitons and related many electron-hole complexes play an important role
in facilitating gain in a ZnSe quantum well, while impurity states coupled strongly to
the lattice appear to be of direct benefit to the GaN LED in spite of a highly defected
crystal microstructure. In the ll-VIs particularly, flexible heterostructure designs have
produced, in addition to the advances with the conventional edge emitting lasers,
recent demonstrations of vertical cavity surface emitting lasers and evidence for very
large normal mode (vacuum Rabi) splittings in a microcavity in the strong coupling
regime. Apart from some fascinating science in photonic nanostructures, these
developments add to the design flexibility in the basic optical/electronic
"infrastructure" for wide gap
optical semiconductors, 1018 n 1 1 1 1 1 1 1 r
hence further increasing the
odds for an eventual IQ"
emergence of
technologically viable blue ^ 1016
diode lasers.
ZnSe (EA=92 tneV)
In the epitaxy of 1015
common ll-VI
heterostructures ZnSe 1014
forms the 'hub' for blue-
green lasers. Lattice
constant mismatch is a 1013
7.0 9.0 13.0
serious design constraint for 1000/T
the ll-VI materials and the
choice of the substrate is not Figure 7 .Temperature dependence of the free hole density
obvious, either. To date, in p-ZnSe:N, p-Zn(S,Se):N, and (Zn.Mg) (S,Se):N (Ref. 6).
305
most light emitters have been grown on GaAs substrates/buffer layers which present
an 0.25% lattice mismatch to ZnSe, but can be fully lattice matched to Zn(S,Se) and
(Zn,Mg)(S,Se) of particular compositions. Results based on ZnSe homoepitaxy are
just beginning to appear. Complicating the epitaxy of the ll-VI ternaries and
quaternaries is the fact that none of the elements have an (even close to) unity
sticking coefficient on the growth surface, in strong contrast with most lll-V cases.
I
/^-Au
(b)
/
D-ZnSe \|>-2nT«
(c)
-600
p-doping in ZnTe:N 0 10 -10 10 -10 10
were achieved, with Voltage (V) Voltage (V) Voltage (V)
free-hole con-
centrations reaching Figure 3: Current-voltage characteristics of a p-ZnSe epitaxial layer,
p=10l9cm-3 [8], and contacted by the ZnSeTe graded bandgap scheme (from Ref. 8).
since e.g. palladium
was found to form a
low-resistance contact to such ZnTe epilayers [9], a 'bandstructure engineered' low
resistance contact became feasible. In particular, a p-Zn(Se.Te) graded bandgap
307
3. Diode Lasers
2 _ f . CD
index
ridge
guided
wave-
I-
O /•'/ ,
O
CO
/// 3 >
guide structures /-'/
to reduce the /// -
■
/''/
injection cur- /// ■
/'/ -
rent. Figure 5 ■
shows the 1
continuous- 6 8 10 12 14
wave output/ Current (mA)
input char-
acteristics at Figure 5: Room temperature, continuous-wave operating characteristics
X=508 nm of of three nominally identical index guided SCH lasers at X = 508 nm.
three nominally
identical 4.2/^m wide devices from the authors' laboratories, where the ZnSeTe
graded bandgap ohmic contacts and high reflectivity facet coatings were also
incorporated [14]. Threshold voltages well below 6V and currents below 10mA have
now been reached, with differential efficiencies up to 60%. Single transverse mode
characteristics have been obtained in the ridge waveguide geometry for both single
and multiple quantum well devices. Initial devices sustained the cw operation less
than a minute before failure but the groups at Sony laboratories and at 3M/Philips
have recently extended the longevity to the scale of hours, illustrative of the rapid
progress in the field (when compared e.g. with the evolution of the GaAs injection
laser in the 1970s). The problem of device degradation and lifetime of the ll-VI lasers
is clearly a vital one, whose solution is imperative for technological applications of
these devices. Given the mechanical properties of these relatively polar
semiconductors (suggesting more vigorous dislocation dynamics), coupled with the
low MBE growth temperature, the control of defects and associate device
degradation poses an inherently larger challenge than in GaAs lasers.
Recently, researchers at 3M, Philips and the authors' group have conducted
systematic studies of the microstructure associated with the device failure [15],[16].
We now know that extrinsic morphological defects forming during epitaxy act as
launching centers for dislocation networks. The defects are predominantly stacking
faults created at the GaAs/ZnSe heterovalent interface, and lead to the presence of
a finite initial threading dislocation density in the strained active QW layer of the laser.
Nonradiative recombination in the QW at nearby point defects induces further
dislocation activity such as the generation of dislocation loops and leads to enetual
optical and electrical degradation of the QW. The dislocation multiplication proceeds
at a finite rate, even under relatively modest current densities (~ 100 A/cm2 as in a
309
LED). Several groups are now focusing their attention to improving the epitaxy of
laser quality material, with specific emphasis on the control of stacking faults and
point defects. For example, the stacking fault density has been reduced from about
10B cm"2 a couple of years ago to about 10* cm2 presently; correspondingly device
lifetime has increased thousandfold.
In parallel to the development of the edge emitting diode lasers, there are
ample reasons to explore other resonator configurations such as the vertical cavity
surface emitting laser (VCSEL). At room temperature the gain of a ZnCdSe QW SCH
diode laser is approximately g^lSOO cm1 (at a current density I-500 A/cm ). In
designing a VCSEL structure, this implies a requirement for the mirror reflectivity of
R=0.997 for a three QW thick gain medium (!_„« 75Ä). While the direct epitaxial
growth of distributed feedback Bragg (DBR) reflectors is in principle possible, the
small differences in the indexes of
refraction e.g. for ZnSSe and (a) SiO /HfOj DBR lUclc
a
The observed Rabi splitting is much larger than obtained in the GaAs
microresonators. While the work with the ll-VI systems is only beginning in this field,
it is possible to envision microresonator arrangements in which Q>kT can be
achieved at room temperature. This implies a very high degree of intermixing of the
electronic and photonic waves and suggests many new and fundamentally distinct
possibilities for blue-green light emitters.
5. Summary Remarks
In this overview we have touched on current work on the new blue-green ll-VI
lasers. The field is moving rapidly and any overview is subject to annual revision.
Clearly, further advances need be made e.g. for improving the vertical transport, the
p-type doping, and QW designs for larger gain. Better control of crystalline defects,
including choices of substrate and buffer layers, and their identification in the device
degradation process are crucial issues now when technological prospects of these
diode lasers need be evaluated realistically and competitively. Much innovation and
basic research is also awaiting the researchers in this expanding field. Extension
deeper into the blue and near UV is being pursued, as well as studies of new lasers
such as small vertical cavity emitters.
References:
[I] M. Haase, J. Qiu, J. DePuydt, and H. Cheng, (1991) Appl. Phys. Lett. 59, 1272
[2] H. Jeon, J. Ding, W. Patterson, A.V. Nurmikko, W. Xie, D.C. Grillo, M. Kobayashi,
and R.L. Gunshor, (1991)>App/. Phys. Lett. 59, 3619.
[3] R.M. Park, M.B. Troffer, C.M.Rouleau, J.M. De Puydt, and M.A. Haase, Appl. Phys.
Lett. 57, 2127 (1990); K. Ohkawa, T. Karasawa, O.Yamazaki, (1991) Jpn J. Appl.
Phys. 30, L152.
[4] D.J. Chadi, (1994) J. Cryst. Growth 138, 295.
[5] D.Laks, CG. Van de Walle, G. Neumark, and S.Pantelides, (1993) Appl. Phys.
Left 63, 1375
[6] J. Han, Y. Fan, M. Ringle, L. He, D. Grillo, R. Gunshor, G. Hua, and N. Otsuka,
(1994) J. Cryst. Growth 138, 464.
[7] J. Han, M. Ringle, Y. Fan, R Gunshor, and A Nurmikko, (1994) Appl. Phys. Lett.
65, 3230
[8] J. Han, T. Stavrinides, M. Kobayashi, M. Hagerott, and A.V. Nurmikko, (1993) Appl.
Phys. Lett. 62, 840.
[9] H. Okuyama, T. Miyajima, Y. Morinaga, F. Hiei, M. Ozawa and K. Akimoto, (1992)
Electr. Lett. 28,1798.
[10] Y. Fan, J. Han, L He, J. Saraie, R.L Gunshor, M. Hagerott, H. Jeon, A.V.
Nurmikko, (1992) Appl. Phys. Lett. 61, 3160.
[II] M. Ozawa, F. Hiel, A. Ishibashi, and K. Akimoto, (1993) Electr. Lett. 29, 503-504.
[12] N. Nakayama, S. Itoh, K. Nakano, H. Okuyama, M. Ozawa, A. Ishibashi, M.
Ikeda, and Y. Mori, (1993) Electr. Lett. 29,1488.
[13] J.M. Gaines, R.R. Drenten, K.W. Haberem, T. Marshall, P. Mensz, and J.
Petruzzello, (1993) Appl. Phys. Lett. 62, 2462-2464.
[14] A. Salokatve, H. Jeon, J. Ding, M. Hovinen, A. Nurmikko, D.C. Grillo, J. Han,
H. Li, R.L. Gunshor, C. Hua, and N. Otsuka, (1993) Electr. Lett. 29, 2192
[15] S.Guha.J.DePuydt, M. Haase, J.Qiu, and H.Cheng, (1993) Appl. Phys. Lett.
63,3107
[16] G.C. Hua, N. Otsuka, D.C. Grillo, Y Fan, M.D. Ringle, R.L. Gunshor, M.
Hovinen, and A. Nurmikko, (1994) Appl. Phys. Lett. 65,1331
[17] J. Ding, T. Ishihara, M. Hagerott, H. Jeon, and A.V. Nurmikko, (1992) Phys.
Rev. Lett. 60, 1707-1710, (1993) Phys. Rev. B47, 10528-10537 .
[18] J. Ding, M. Hagerott, P. Kelkar, A.V. Nurmikko, D.C. Grillo, Li He, J. Han, and
R.L. Gunshor, (1994) Phys. Rev. B50, 5787.
[19] OH. Henry, R.A. Logan, and F.R. Merritt, (1980) J. Appl. Phys. 51, 3042.
[20] H. Jeon, V. Kozlov, P. Kelkar, A.V. Nurmikko, D.Grillo, J. Han, M. Ringle, and
R.L. Gunshor, (1995) Electronics Lett. 31,106; (1995) Appl. Phys. Lett. 67,1668
[21] P. Kelkar, V Kozlov, H. Jeon, A. V. Nurmikko,C.-C. Chu, D. C. Grillo, J. Han,
C.G. Hua, and R. L. Gunshor, (1995) Phys.Rev.B52, R5491
ORGANIC TRANSISTORS — PRESENT AND FUTURE
G. HOROWITZ
Laboratoire des Matöriaux MoUculaires
C. N. R. S., 2 rue Henry-Dunant
94320 Thiais
France
1. Introduction
Organic materials are almost everywhere in electronic devices. They are used for instance
in lithography and encapsulation. They are everywhere, but at the very heart of the
device, upon which silicon still imposes its dictatorship. Nevertheless, organic
semiconductors do exist, and have indeed been largely studied since the early fifties [1].
It has been shown that metal-semiconductor (MS) and metal-insulator-semiconductor
(MIS) structures can be realized with these organic materials. Practical applications were
even thought to be within reach in 1978, when a photovoltaic cell made with
merocyanine — an organic dye — was claimed to present a power efficiency close to 1
% under AMI solar illumination [2]. Unfortunately, this yield, which is still one order
of magnitude too small, has not been improved to date, but the domain is still active [3-
5].
The second chance for organic semiconductors came in the early eighties, with the
emergence of conjugated polymers and oligomers. Conjugated polymers present the
unique property of having their conductivity increased by several orders of magnitude
upon doping. Highly conducting polyacetylene has been claimed to present a
conductivity close to that of metals [6]. It has been recognized more recently that the
non intentionally doped ("undoped") form of conjugated polymers and oligomers
constitute a new class of organic semiconductors. A tremendous amount of work has
been done on light-emitting diodes made with organic conjugated polymers, [7-9] after
it appeared that these devices could compete with their mineral counter-part in term of
efficiency and extent of colors that can be produced.
The field-effect transistor (FET) is another device where conjugated polymers and
oligomers have proved useful [10-16]. Of these, polythiophenes [10,12,14,16] and
oligothiophenes [13-15] seem to be the most promising. Polythiophenes present a field-
effect mobility ranging from 10"5 to 10"4 cm2V"1s"1, whereas that of oligothiophenes,
and more precisely sexithiophene (6T) and some of its derivatives, is two orders of
magnitude higher. Although these values are far below those of mineral semiconductors,
they are approaching that of hydrogenated amorphous silicon (a-Si:H), which would
make organic materials particularly suitable for large area devices, e.g., flat panel
displays. The electrical properties of these polymers and oligomers have been found to
315
S. Luryi et dl. (eds.), Future Trends in Microelectronics, 315-326.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
316
be much dependent on the purity of the material and method of film preparation [17,18].
They can also be modulated through chemical derivations.
In the present paper, we describe the fabrication and operating mode of organic field-
effect transistors (OFETs), and review devices published to date. We will then focus on
oligothiophenes, and show that a strong correlation can be found between their electrical
and structural properties. Accordingly, the device performance can be optimized by both
chemical and physical means.
Figure 1 gives a schematic view of the most commonly used structure. OFETs are
fabricated according to the thin film transistor (TFT) architecture, which was first
described thirty years ago [19]. Because an insulating layer is not easy to deposit on
top of an organic semiconductor, the structure is inverted, i.e., it is built over the gate
electrode. Most often, the insulator is silicon oxide thermally grown on a silicon wafer.
The Si substrate, which plays no active role in the device, serves as the gate electrode.
Two gold source and drain electrodes are evaporated on the silicon oxide before the
deposition of the active organic semiconducting layer. They form ohmic contacts to the
semiconducting layer.
Vd
-10 -
-5 -
The device can operate either in the enrichment mode, Figure 2a, or in the depletion
mode, Figure 2b. The signs of both Vj and Vg are consistent with a p-type
semiconductor (here 6T). In the enrichment mode, the curves can be described by
Equations (1) [22].
W
Id.sai = 2L C' WET (Vg-VtY (Vd>Vg) (lb)
318
Here, W and L are the channel width and length, respectively, C; is the insulator
capacitance (per unit area), ßFET is the field-effect mobility and Vt the threshold
voltage. Curves in the enrichment mode can thus be used to determine the field-effect
mobility.
In the depletion mode, the curves show that the semiconducting layer can be fully
depleted. (The small remaining ohmic current is due to leakage through the insulator.)
This occurs when the width of the depletion layer equals that of the semiconducting
layer, that is when the gate voltage equals the pinch-off voltage Vp, Equation (2)
[22,23].
Here, q is the electron charge, N the dopant concentration, d the thickness of the
semiconductor, es its dielectric constant, SQ the permittivity of free space, and Cs the
dielectric capacitance of the semiconducting layer (Cs=esEQ/d). Equation (2) can be
used to estimate the doping level of the organic semiconductor.
The organic materials used in field-effect transistors can be sorted into two main groups:
polymers and molecular materials.
■J^N^N^S-. Trans-polyacetylene
n
4*h n
Polythiophene
CLH
mn2m+l
^k (34. Poly(3-alkylthiophene)
<*M. n
Poly(thienylene-vinylene)
By this we mean materials made of smaller molecules than polymers. They comprise
conjugated oligomers, and other conjugated molecules. We note that the early organic
semiconductors, e.g., anthracene, were members of this class of materials.
C
nH2n+l
ss s s
^\f \J \i
CnH2n+l
Figure 4. Molecular structure of sexithiophene (top) and its derivatives substituted with
alkyl chains at end (middle) and pendent (bottom) positions.
4.2. OLIGOTHIOPHENES
Oligothiophenes deviate from the general trend observed above. This is probably
because of their self-assembling behavior, which leads to a crystalline structure of the
evaporated films.
also note that the conductivity of «Ts is anisotropic, the charge transport being favored
in the direction parallel to the film.
We shall now focus on 6T and its derivatives. We have recently resolved the crystal
structure of non substituted 6T. The unit cell, Figure 5, is monoclinic, and contains four
molecules arranged in a herringbone close-packing [35].
This crystal arrangement presents some remarkable features. Molecules are rigorously
planar and strictly parallel to each other. Moreover, the n-orbital overlap between nearest
neighbor molecules is maximum, which can account for that, unlike conjugated polymer,
the charge transport is favored along the rc-stack direction.
The crystal structure of sexithiophene substituted by pendent butyl groups has been
resolved recently [36]. The main effect of the pendent group is to shift the nearest
neighbor thiophene chains, which results in a dramatic decrease of the Tc-overlap. This
would account for the very low conductivity and mobility of that compound.
End substituted oligothiophene present a two dimensional structure, where layers of
thiophene chains alternate with alkyl layers. Up to the diethyl-quinquethiophene
(DE5T), the unit cell has been claimed to be orthorombic [37]. In spite of great efforts,
we have not been able to grow single crystals of dihexyl-sexithiophene (DH6T), in
which both the thiophene and alkyl chains are longer. However, X-ray diffraction on
evaporated films of DH6T showed a two dimensional structure extending over large
distances, that would compare to that of a liquid crystal.
multiple thermal trapping and release model [38]. Data are consistent with an
exponential distribution of traps, that compares to that found in a-Si:H. The traps are
attributed to grain boundaries, and the higher mobility of the end-substituted DH6T can
be ascribed to its particular liquid-crystal-like structure, with a very low density of grain
boundaries in the direction parallel to the film. Our model can also account for the
enhanced mobility of 6T vacuum evaporated on heated substrates, in which electron
microscopy showed an increase of the crystal grain size [39].
The deposition on a heated substrate also results in a decrease of the conductivity,
which can be attributed to an improvement of the purity of the film. An even lower
conductivity was measured on single crystals. In that case, the mobility is close to 0.1
cn^V'V1, and the doping level, estimated from the pinch-off potential, is 3xl014 cm"3,
which corresponds to a molar ration of 0.2 ppm. Table 5 compares the electrical
parameters of polythiophene to that of sexithiophene in its different forms.
TABLE 5. Conductivity, mobility and doping level of polythiophene and different forms of
sexithiophene
Material Conductivity Mobility Doping level
(S/cm) (cnvty-V1) (cm"3)
7 6
Polythiophene 10" - 10" 10"5 - 10"4 1017- 1018
6
6T (substrate at 25°C) 1.5xl0- 0.003 lxlO18
6T (substrate at 280°C) 1.2xl0"7 0.025 5xl016
6T (single crystal) <10"9 0.075 3xl014
DH6T 6xl0"5 0.04 3xl017
These results indicate that, in contrast to "amorphous" organic compounds, large on-off
ratios can be obtained with sexithiophene.
7. Conclusions
Field-effect transistors have been made with thin films of a number of organic
semiconductors, including conjugated polymers and oligomers. Two categories of
behavior can be differentiated. First, in most of the conjugated polymers, and in a great
number of "amorphous" molecular materials, conduction is governed by a hopping
mechanism. This result in a field-effect mobility that depends on the doping level.
Accordingly, high mobility is only obtained with materials that also present a high
conductivity. FETs made with these materials present an inherently poor on-off ratio.
The second category is that followed by the oligothiophenes. In these materials, the
charge transport obeys a thermal trapping mechanism, and is hence only dependent on
the density of traps, whereas the conductivity depends on the doping level. High
mobility can thus be associated with low conductivity by using highly ordered and very
pure materials. Ordering can be obtained by chemical means, as it is favored by
substitution at ends of the molecule. Highly crystalline films can also be realized by
adjusting the deposition parameters.
324
Practical applications of the organic EETs can be envisioned in fields where large areas
are needed. Another interesting property of these devices is the possibility to make them
on flexible substrates. We finally note that the fabrication of OFETs requires soft
techniques and very low temperatures.
8. References
Abstract
1. Introduction
The development of III-V epitaxial growth techniques such as molecular beam epitaxy
(MBE) and metal-organic chemical vapor deposition (MOCVD) has made possible the
design and reproducible growth of a broad range of novel optoelectronic and high speed
electronic devices based on ultra-thin layers. High electron mobility transistors and
heterojunction bipolar transistors offer substantial improvements over traditional devices
[1]. Other applications of heterostructure growth provide new types of devices requiring
ultrathin layers, such as quantum wells for lasers, resonant tunneling diodes, etc. In
much of this work the thin layered structures are designed to confine carriers, or to
realize thin potential barriers to take advantage of quantum tunneling effects. Among the
advantages of high growth control are the ability to do modulation doping and delta
doping, with control on the scale of a monolayer [2].
Semiconductor quantum wells (QW's), achievable through control of layer
thickness with monolayer accuracy, have been a rich source of insight into
327
S. Luryi et al. (eis.), Future Trends in Microelectronics, 327-335.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
328
semiconductor physics and have led to novel structures for electronic and photonic
device applications. The interaction between photons and carriers in a quantum well, the
optical transition, is governed by the fact that only discrete energy states are allowed for
electrons and holes in the well. Not only do quantum wells provide transition energies
different from the bulk material, but also population inversion is achieved at a lower
threshold current in a QW laser. Confinement of carriers can be substantial in 8-doped
quantum wells. For example, we have reported confinement of holes on a scale of about
5 A in a 50A QW 8-doped with Be [2]. Carrier confinement can lead to a number of
interesting low-dimensional effects, including those related to quantum wires and
quantum boxes. The quantum-confined Stark effect is an example of a useful effect (for
light modulators) that can only be achieved with these modern growth techniques. In the
usual quantum well structure, where electrons and holes are well-confined in the
potential well region, transition energies occur near the energy gap of the well layer.
For applications requiring higher transition energies, a number of quantum well
structures have been proposed and synthesized, such as those using narrow wells, high-
gap well layers, and superlattices in the well region. From the earliest days of MBE,
there has been an interest in artificial periodicities available by growth of multilayer
heterostructures. Both compositional and doping periodicities can lead to new
"miniband" or "subband" conduction of electrons and holes. Superlattice quantum wells
(SLQW's) can be used to achieve high-energy transitions, which can be varied over a
few hundred meV using different Al As and GaAs layer thicknesses in the SLQW's. It is
also possible to use aperiodic layer thicknesses, including random-period superlattices,
to further tailor the properties of the quantum wells [3].
3. Photodiodes
Recently we have applied the Bragg reflectors available by MBE growth to take
advantage of resonant absorption of photons in microcavities. Such resonant cavities
can have enormous impact on traditional photodiodes.
The PIN photodiode is the most widely deployed photodetector for photonic
applications. The light that enters the photodiode is attenuated exponentially with
329
distance into the absorbing layer, and photogenerated electrons and holes give rise to a
photocurrent that is proportional to the incident intensity. There is a tradeoff between
the responsivity and the bandwidth of the PIN structure, since to achieve high quantum
efficiency a relatively thick absorption layer is required, which in turn requires a longer
time to collect the photogenerated carriers. This is the origin of the transit time limit to
the bandwidth.
We have demonstrated a novel resonant-cavity photodiode, shown in Fig. 1, that
circumvents the quantum efficiency bandwidth tradeoff due to transit time effects [6].
This structure increases the absorption through multiple reflections between two parallel
mirrors in a Fabry-Perot cavity whose length is typically one wavelength. The lower
mirror is an integrated Bragg reflector consisting of alternating X/4 epitaxial layers,
having a reflectivity >99%. The top mirror is usually a high reflectivity dielectric stack
that can be deposited after fabrication and initial characterization.
.Dielectric Stack
-■I j
v ■ "-wtwr
Absorbing
Region
L2 I
} X/4DBR
Ivvvvvvwvvvvv^^11 Contact
Figure 1. A resonant cavity PIN photodiode with distributed Bragg reflectors to define a
cavity one wavelength long.
For resonant-cavity photodiodes having device areas -50 \lvcfl, bandwidths greater
than 100 GHz can be achieved without sacrificing responsivity. The photon buildup
time in such a short cavity corresponds to a bandwidth in excess of -lO1^ Hz and poses
no limitation on the speed. This illustrates one advantage of the resonant-cavity
approach, namely, that the quantum efficiency can be effectively decoupled from the
transit-time [6]. While the improved bandwidth is accomplished at the expense of a
narrower spectral response, almost all photonic systems operate within a very narrow
wavelength range. In fact, this may be used to advantage for applications such as
wavelength-division multiplexing in fiber optic systems. Resonant-cavity structures
promise performance enhancements in several photodetector applications [7].
330
It is well known that the internal gain of avalanche photodiodes (APD's) can provide
substantial improvements in signal-to-noise compared to PIN photodiodes. The gain is
achieved through carrier multiplication, a consequence of impact ionization at high
electric fields. Two of the crucial performance characteristics of APD's, the gain-
bandwidth product and the excess noise arising from the random nature of the
multiplication process, are determined primarily by the electron and hole ionization
coefficients (a and ß, respectively) or, more specifically, by k which is defined as the
ratio of the ionization coefficients. For low noise and high gain-bandwidth products, a
large difference in the ionization rates (k « 1) is desirable [8].
In a properly designed APD only the carrier with the highest ionization rate (either
the electrons or holes, depending on the material) is injected into the high-field
multiplication region. This can be accomplished with separate absorption and
multiplication (SAM) structures. The SAM APD's have been widely deployed in long-
wavelength, high-bit-rate optical transmission systems. As these transmission systems
have progressed to higher and higher bit rates, however, the bandwidth of the SAM
APD's has become a limitation. Therefore, the resonant-cavity structure has become an
attractive alternative to achieve high speed without sacrificing quantum-efficiency.
Recently, we have successfully incorporated a SAM APD into a resonant-cavity
structure [9]. A schematic cross section of the device is shown in Fig. 2. It has a 500 A
2000Aptype(1017cm"3)
A!Q 4GaQ gAs multiplication layer
500 A p'doped Ing [Gag gAs absorption layer
The vertical-cavity surface-emitting laser has recently gained much attention in the
research literature. This type of structure represents a minimum volume laser, since the
Fabry-Perot cavity length is on the order of the lasing wavelength, and thus it has the
greatest potential of any semiconductor laser structure for ultra-low threshold current.
The device's vertical geometry is tailor-made for large scale integration, since light is
emitted normal to the epitaxial surface, and thus is also compatible with large area two-
dimensional operation of phased arrays for high power operation. From a manufacturing
standpoint, wafer scale testing is possible with the vertical-cavity laser. This is a major
advantage over the traditional edge emitter, which must be cleaved and mounted for
testing.
A common problem with VCSEL structures is that current is injected through the
same region from which the light is emitted, unlike edge emitters in which current and
light emission are along different directions. Thus, optimizing the top Bragg reflector
mirror has in the past led to high series resistance for the current, and a resulting high
bias voltage at the threshold for lasing. Current funneling may be achieved through the
use of MBE regrowth over a patterned n-type current blocking layer placed within the p-
type top mirror near the active region. However, these thin n-type layers pass a
considerable amount of leakage current. In addressing this problem, we began by
incorporating a layer of AlGaAs grown at a reduced temperature as a semi-insulating
region to allow current funneling from the top contact to the active region of the laser
[12]. This AlGaAs region is lattice-matched to the rest of the structure, and we have
developed methods for selective etching and regrowth to provide the buried semi-
insulating current funneling region in these devices. Post growth processing consists of
a Cr/Au metallization and deposition of a quarter-wave ZnSe/CaF2 top DBR.
Commercial VCSEL fabrication is based largely upon the process of proton
implantation. The proton implantation is used to damage the epitaxial semiconductor
332
crystal below its surface, greatly decreasing its conductivity by trapping charge carriers,
while leaving an electrically conducting path near the crystal surface. By performing the
implantation into selected regions of the crystal, electrical current can be funneled into
the VCSEL active region, thus exciting the electron-hole pair density necessary to
achieve lasing. This proton implantation process is a carry over from an older edge-
emitting laser diode technology, but one which is relatively cheap, well-characterized,
and is adequate in realizing reasonable performance VCSEL's. In both the buried semi-
insulating AlGaAs method and the proton bombardment, electrical isolation is achieved
with little change in refractive index. Therefore, lateral optical confinement requires
another approach.
In 1990 Holonyak and co-workers at the University of Illinois noted that a single crystal
film of AIAs could be selectively converted to high quality AlxOy using "wet"
oxidation at about 400°C [13]. In January of 1994 it was demonstrated by Deppe and co-
workers at UT-Austin that the selective oxidation process can produce very low
threshold VCSEL's [14,15]. The benefit of this process in VCSEL fabrication is two-
fold: it serves as a buried insulator to tightly confine the injected current to a small area
active region, and it provides a lateral index-guide for optical confinement.
Figure 3 (a) shows a schematic cross-section of a full-wave cavity layered
structure, and Fig. 3 (b) shows a scanning electron microscope (SEM) image looking
down on a selectively oxidized epitaxial heterostructure of AlAs-GaAs InGaAs [14]. In
Fig. 3 (b) the AlxOy oxidation front has diffused laterally under a GaAs mesa (see Fig.
3 (a)) leaving a 4 |lm square AIAs region which becomes the active part of a fully
processed device. For the selectively oxidized VCSEL the major advantage of the half-
wave cavity [16] as compared to the full-wave cavity [14] is the ability to place the
AlxOy layer close to the light emitting quantum well region and the lasing mode
intensity peak (at the center of the vertical-cavity). The half-wave cavity, therefore, is an
important design feature for maximizing optical index confinement, and therefore
minimizing threshold.
The selectively oxidized half-wave cavity VCSEL is unique in that it is an index-
guided laser which requires no epitaxial regrowth step and yet provides strong optical
confinement as well as excellent current confinement. Our research group was first to
demonstrate the application of the native oxide process to VCSEL fabrication [14], and
recently we have demonstrated that the use of a half-wave cavity spacer can lead to
extremely low threshold currents [16]. Our recent work has used a combination of
epitaxially grown AlAs/GaAs and post-growth deposited dielectric CaF2/ZnSe Bragg
reflectors. The result due to the optical and electrical confinement has been the lowest
threshold VCSEL's yet realized, with room-temperature threshold currents under 100 |iA
and a record low threshold current of 59 |lA at an optimized temperature of 250K [16].
Besides the more fundamental interest in studying ultra-small cavity lasers and exploring
the limits of minimizing threshold for a semiconductor laser, it is likely that the
important area of digital signal transmission will benefit in terms of switching speed by
reducing the laser diode threshold current well into the sub-100 (J.A regime [17]. Present
commercial VCSEL technology based on proton implantation confinement is limited to
threshold drive currents in the 2-5 mA range.
333
(a)
p-GaAs
Al
/ p-AIAs *J x°y
InGaAs/GaAs
:i< i-AIAs/GaAs
DBR
-n-GaAsSub.
Figure 3. (a) Schematic cross section of the buried ring contact VCSEL structure
showing the role of the lateral oxidation of the AlAs underneath the GaAs mesa
defining the device active region [14]. (b) SEM photograph of a native-oxide
defined 4 p,m square AlAs region buried beneath a 30 pm diameter GaAs mesa.
The oxide layer surrounding and beneath the GaAs mesa provides
device isolation for current injection [14].
Since the initial demonstration in 1994, first reported as a late paper at the Conference
on Lasers and Electro-Optics [15], the impact of selective oxidation on VCSEL research
has been impressive [14-16, 18-20] and has led to several record results [16, 18, 19].
The native-oxide defined VCSEL has yielded the lowest threshold currents (sub-100 \iA)
yet achieved in a laser [16, 19], as well as the highest power conversion (wall-plug)
efficiency of >50% [18]. For comparison, typical threshold currents of proton implanted
VCSEL's are in the 2-5 mA range as mentioned above, with the highest reported wall-
plug efficiency being -20% [18]. An important demonstration was made recently by
Sandia workers who showed device reliability of over 28 days of continual operation
with little degradation for an all epitaxial structure [21], giving strong evidence that the
AlxOy defined VCSEL can be made to operate reliably.
Clearly, the use of microcavity devices employing DBR's will play an important role in
the emitters and detectors of the future. Furthermore, we have enough evidence to
believe that lateral confinement of light will likewise contribute significantly to vertical
cavity laser structures. An intriguing question remains regarding what new physics will
emerge from the very recent ability to "confine" light in three dimensions on the scale
of a wavelength. The confinement vertically is quite good, and can be easily
dimensioned to a wavelength or a fraction such as X/2. In the lateral dimension, the
recent use of selective oxidation allows us to define the cavity within a few wavelengths
334
and in the future it may be possible to choose lateral cavity dimensions with even more
accuracy. Obviously, the refractive index change between the semiconductor and AlxOy
does not lead to complete confinement. Therefore, an analogy with the confinement of
carriers is not accurate. On the other hand, this is an example of a new regime for
experiments that did not exist before, and one expects surprises when such things
happen.
8. Acknowledgments
This work was supported by the Joint Services Electronics Program under contract
F49620-92-C0027, and by the Texas Advanced Research and Technology Programs.
9. Bibliography
J-M. P. DELAVAUX
AT&T Bell Laboratories, 9999 Hamilton Boulevard., Breinigsville,
PA 18031, USA.
1. Introduction
An integrated laser system (as in an optical fibre laser system) obtains a major
advantage from the significant spatial overlap of the pump radiation mode with that of
the output radiation mode, as they propagate in a waveguide. Consequently two factors
337
S. Luryi et al. (eds.), Future Trends in Microelectronics, 337-351.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
338
Interferometric
Laser source sensor (MZ)
dbg dbg
o
PC
zr
external
effect
Legend:
Ring laser
p: diode pump
c: coupler
f: fan-out
dbg: distributed Bragg
grating miror
wdm: wavelength division
multiplexor
d: detector
are in play, the spectroscopy of the active ions in the host medium, in this case the
lanthanides or rare-earths in a glass host, and the opto-geometrical properties of the
waveguide, i.e., its numerical aperture and modal diameters. Our goal was therefore, to
make a low-loss micro-waveguide in a glass host containing a small percentage of rare-
earth, whilst maintaining the desired spectroscopic properties. To this end we have
spent several years in finding the correct glass host and the optimum ion-exchange
conditions.
The ion exchange process itself consists of immersing a glass wafer, the desired form of
the wave-guide suitably defined by a photo-lithography, into a molten bath of an
339
alkaline salt. The unprotected part of the glass is then subject to a migration of ions,
for example Na <-» K or Ag, and a channel of higher optical refractive index (= 1 %) is
produced. A second stage effectively buries the guid
at a depth of a few micrometres below the surface to reduc
losses. This process was pioneered at the Institut National Polytechnique de Grenoble
[1] and in Japan [la].
We have found that the ion exchange process does not change the spectroscopic
properties from that of the bulk material to any significant degree."
We have chosen two rare-earth laser systems, neodymium, and erbium co-doped with
ytterbium.
Neodymium has a strong transition at 1055 nm which terminates at a level above
the ground state. The erbium transition at 1530 nm terminates at the ground state and
therefore has an absorption band which must be overcome before net gain is reached.
On the other hand the lifetime of the excited metastable state is much longer in
erbium making it a strong candidate for stimulated emission and in addition the erbium
transition lies in the 1530 nm band, an optimum propagation window in optical fibre
and therefore widely used in the telecommunications industry.
The absorption bands vary only slightly with different glass hosts, but more
significantly, the radiative time constants can vary considerably depending on the
phonon energy spectrum of the glass host. This is more advantageous in some glasses
and we have chosen to use a phosphate oxide glass as distinct from a silicate glass.
A necessary condition for exciting the active ion is a pump absorption band which
corresponds to the wavelength of an available powerful semi-conductor laser diode. In
neodymium there is a strong absorption at 800 nm which we exploit. In erbium the
preferred pump band is at 980 nm but the absorption is not strong so we have co-doped
it with ytterbium, which has a stronger band in this region. In this co-doped system the
ytterbium ion is excited (pumped) and its energy is transferred to the erbium ion. A
shorter waveguide will then be the optimum.
2.3 MODELLING
The complex marriage between the spectroscopic properties and the waveguide geometry
may be modelled to some degree and eventually one can calculate the optima for the
performance of a device. We can also model second order parasitic phenomena and
thereby seek to improve the concentrations of the basic materials.
Our mathematical model consists of solving the coupled differential equations of
propagation, having first determined the steady state carrier densities at each significant
energy level by solving the rate equations. As the optical powers in the guide are also
radially dependent, there is an additional requirement to solve an intensity dependent
spatial integral.
A comparison of the model and the experimental results has shown that second
order phenomena, probably due to co-operative energy transfer between dopant ions, are
important and that significant improvement in performance may be expected by
optimising the doping and the geometry of the amplifiers and lasers.
340
2.4 OPTICAL AMPLIFICATION AND LASING AT 1054 nm [2]
(1)
^k = Rf Nf - Amf.Nm - Wmf.Nm - C.NZ
df
where Rf,.Nf is the absorption term of the ions in their fundamental level, Amf.Nm and
Wmf.Nm are respectively the spontaneous and stimulated emissions from the metastable
level. C is an up-conversion coefficient which is independent of the rare-earth
concentration. 23
For C = 10" mV, we obtain a reasonable agreement between the experimental
and theoretical curves. On eliminating up-conversion (C = 0) the model predicts a gain
of 15 dB for the same pump power.
We have adapted our ion-exchange process, using silver ions, to our phosphate glass
doped with 2% by weight of erbium and 4% by weight of ytterbium. This gives us the
possibility of making short 44 mm microguides, single mode at both pump and signal
wavelengths, 980 nm and 1540 nm, respectively. The average mode diameter (@ 1/e) of
341
these guides for the two wavelengths are respectively 4.7 fim at 980 nm and 7.5 \im at
1540 nm. These guides are buried 3.5 (im below the surface to avoid surface losses and
enable high coupling efficiency with fibres of up to 89 %. We have measured the net
gain obtained from these structures operating as travelling wave amplifiers. The small
signal net gain at 1537 nm for a pump power of 100 mW was 11 dB for a double pass
and 6 dB for a single pass. Large signal tests indicated that the -3 dB saturation point
occured at an output power of 10 dBm.
An optimisation of the waveguide diameter and the pump" wavelength, together
with an optimisation of the rare-earth concentration, will allow us to make efficient
optical amplifiers on compact glass substrate of less than 2.5 cm2.
2.6.1 Introduction
We report here our first results of an integrated optical laser made by ion-exchange in
Er:Yb phosphate glass. The structures show good slope efficiencies in double pump
configurations and tunability when used with an external fibre bragg reflector.
! 4h Slope=16%
ä
u
2 -
o
&. 1 -
3
o Threshold=67mW
*JL ' . i—
50 60 70 80 90 100
Having discussed the signal generation and amplification part, we now move onto the
signal retreving part by WDM technique which is particularly important for enhancing
the transmission capacity of fiber optic transmission lines. Specifically, we present our
work on new WDM devices in doped glass.
The ideal characteristics of a WDM device include a high wavelength selectivity, a
large channel fan-out and a low loss. In the meantime, a base technology that allows
good fiber connection (or low insertion loss) and is compatible with the other elements
of the transmission line, i.e., laser sources, modulators, amplifiers and photodetectors is
also an important consideration. In attempts to achieve some of these ideal
characteristics, various configurations of integrated WDMs have been designed and
demonstrated on various technologies. The curve diffraction grating WDM, the Mach-
Zehnder WDM and the arrayed-waveguide WDM are among the most studied.
The curve diffraction grating (or echelle grating) WDM [5] is an integrated copy of
conventional bulk-type spectrometer. It is well suited for numerous wavelength
channels with relatively large wavelength spacing (a few nm). Its main weak point is
the high insertion loss, typically of about -15 dB. The Mach-Zehnder WDM [6] uses
the wavelength selective coupling property of a 2-wave interferometer. It is a good
candidate for very small channel spacing of a few A. Its drawback is the maximum
fanout of only two channels. To obtain larger fanout, it is then necessary to cascade
them which increases the size and the losses of the device. The arrayed-waveguide WDM
[7] is in fact a hybrid of the two previous kinds of WDMs. The diffraction grating is
replaced by the array of channel waveguides that acts like a multi-wave Mach-Zehnder
interferometer. .
In search of an alternative WDM device technology that is relatively simple,
compact, high performance and is potentially implementable and integratable with the
amplification and lasing elements discussed in the previous sections, we have
theoretically and experimentally explored a totally different type of WDM
concept—superimposed gratings.
The first successful implementation of multiple sets of gratings was demonstrated
343
The basic concept of a supergrating WDM with multiple sets of gratings superimposed
on a planar waveguide is not difficult to understand. The base structure comprises a
planar waveguide, made of dielectric or semiconductor films, and multiple sets of Bragg
gratings photo-inscribed or etched into a portion of the guide. Each set of gratings is
designed to diffract one wavelength in one direction for a common input angle. To carry
out the actual device design, we should answer several questions pertinent to the
superposition of multiple gratings.
Would the light of a particular wavelength get bounced back and forth by all the
gratings, "see" the combined effect and come out in a way other than that is expected of
a single set of gratings? If yes, what would be the combined effects on the diffraction
direction? efficiency? and wavelength selectivity? It is not a trivial task to answer these
questions because of the possible couplings between the gratings. To study this
coupling between superimposed gratings we have developed a theoretical (coupled-mode)
model [10-12]. With this model, we were able to show that there indeed could be
significant couplings between gratings, that the effects of the couplings could greatly
affect all aspects of the diffraction characteristics, and that it is possible to find WDM
configurations for which the gratings are quasi-decoupled. In regimes of very weak
coupling, the effects of the other gratings, co-existing with the grating designed to
diffract a particular wavelength, are that of largely non-coherent scatterings off
distributed "centers" of small index perturbations. Such effect manifests itself mainly as
a rise in the background "noise" level, namely, a background loss to the efficiency of
diffraction of a given light. In these weak coupling regimes, the task of designing a
supergrating WDM device is greatly simplified -one can essentially design each set of
grating as if the other gratings were absent except for an overall efficiency loss. Of
course, this is limited to relatively large wavelength and angular spacings (i.e., a
relatively small number of channels).
3.2 MODELLING
The Bragg and near Bragg angle of incidence, as well as the multiple-scattering in
successively recorded volume holographic gratings have been theoretically investigated
[13]. In most of the cases the coupled mode theory was used. Although the results
obtained are useful for us, these models are not suitable for the case of multiple gratings
344
in planar waveguides, particularly for WDM applications. In the case of multiple
volume holographic gratings for data storage applications, the light beams used to write
and read-out the gratings are the same, they travel perpendicularly to the recording media
and are at a fixed wavelength (the effect of multi-gratings on the wavelength selectivity
of each of them was not studied), whereas in our case the gratings are read-out by a
multi-wavelength beam travelling within the guide itself. By extending the treatment
for the single grating in guiding structures [14] to the cases of multiple gratings and by
including the couplings, we developed a model for superimposed gratings in planar
waveguide [10-12]. Further improvements to the model have been made and will be
presented in detail elsewhere.
The model allows us to compute the wavelength and angular selectivities and
diffraction efficiencies of multiple (and possibly closely spaced) wavelengths by
multiple gratings either inscribed in the volume or etched on the surface of the guide.
The gratings are characterized by their orientation, their periodicity and their strength
which is calculated taking into account the properties of the guided propagation
(polarization and mode confinement). Mathematically, the coupled differential equations
of propagation that describe the energy exchange between the input beam and the
diffracted ones are numerically solved over the interaction (grating window) area.
where AX(I is the wavelength separation between the adjacent channels and AX0 is the
total width at the first two zeros of the isolated grating spectrum, (i.e. calculated as if
the grating was the only one present in the guide. See inset of Figure 3).
In the" quasi-decoupled regime, the analysis (and design) for a /V-channel WDM is
relatively straightforward. For example, from [15] we can extract the expression for
AX0 as function of the key structure and operation parameters. Assuming a perfectly
collimated input beam, we have:
AAo = A cos(6j) (3a)
Xd 1 -cos(örf- 0,-)
345
: ^3" ^ (3b)
with ne Lc
where Lc is the so-called "coupling length" of the (single) grating, ne the guide effective
index, 0, the input beam incidence angle, Xd the diffracted wavelength and 6d the angle
of diffraction. This equation shows that grating selectivity depends on the angle of
diffraction. As shown in Figure 3, it increases (AX0 diminishes) when 9d increases.
30 I i
Ldiffract on efficient cy
20
\Ai/
J
'< I
\:A/. AX
A.
c
•<3
10
0 L
20 40 60 80
0d (deg)
Figure 3: Wavelength selectivity of a single grating versus diffraction angle (0, = O).
Now, if we consider the design of a WDM with evenly spaced channels of the
quantity A\WDM one can see there is a minimal diffraction angle under which the
gratings will be coupled and will present a distorted spectrum. This minimal diffraction
angle, 8dmin, is so that Al0 (9dmin) = AXWDM. For 0,- = 0, which is the most convenient
configuration, one finds:
For a central wavelength Xd, a channel spacing AXWDM and guiding structure given by
the effective index, 6dmin will depend on the Lc of the grating. The longer the grating,
the smaller the minimal diffraction angle. For example for Xd = 1.3 (im, AXWDM = 2
nm and ne = 1.5 we calculate 8dlnin = 66, 39 and 29 deg for Lc = 1, 5 and 10 mm (see
Figure 4).
346
Figure 4: Minimal diffraction angle versus grating length for different values
of WDM wavelength spacing. Calculation parameters: 0, = 0, hi= 1.3 um and
n,= 1.5.
These considerations about the single grating wavelength selectivity show that
both angular ranges [0dmin, rc/2] and {-6dmin, -K/2] can be used for a superimposed
gratings WDM to double the fanout capacity.
LC = K Vcos(ftf) (5)
K
2
For example when 6d varies in the range 30-80 deg, Vcos(0J varies within a factor of 2.
Then if we want to use the entire window of diffraction angles, the strength or the
len°th of each grating will have to be adjusted independently to match Lc.
°The second condition on which the effective diffraction efficiency depends is related
to the overlap between the divergence of the input beam and the grating angular
selectivity. Only the part of the input beam inside the cone of acceptance of the grating
will be diffracted. When limited by diffraction the divergence of the input beam is
typically of a few milliradians, depending on its width and the wavelength. On the other
hand the grating angular selectivity is given by (see Figure 5):
347
cos(6j)
AOio = A (6)
sin(&/-0,)
For the numerical values used previously and for a beam width of 1 or 2 mm, we find
that the angular selectivity is one order of magnitude smaller than the input beam
divergence, that is an equivalent loss of about 10 dB. One way to limit the input beam
divergence, thus reduce the loss, is to enlarge its width by inserting a large angle taper
between the input channel waveguide and the grating area. Another way is to not use a
collimated beam but to focus it on the grating area instead. In such a configuration one
takes advantage of the fact that at the focus point a gaussian beam is a plane wave. The
beam width and the focal length used will then have to be carefully calculated.
öd (deg)
In this section we report for the first time the realization of a 4-channel superimposed
»ratings WDM on photosensitive Ge-doped silica planar waveguide on silicon. The
aiding structure is shown on the inset in Figure 4. Only the Ge doped guiding layer is
photosensitive at 250 nm wavelength. The implementation process starts first with the
high pressure hydrogenation of the sample (1500 PSI at room temperature for a couple
ofdays) to increase its photosensitivity. Then the gratings are successively written with
a KrF excimer laser holographic writing set-up similar to the one described in [17]. The
four superimposed gratings have been designed to diffract A,- = 830, 840, 850 and
860 nm at 9di = +32, -26, -30 and +28 deg respectively. The choice of the diffracted
wavelengths has been dictated by the range of tunability of the laser used for the
characterization. Of course, the design can readily be transposed to the useful
telecommunication wavelength windows at 1.3 (im or 1.5 Jim. The gratings common
length is 5 mm. They have been exposed for 15 mn with a -1.8 cm2 laser beam of
-100 mJ/pulse at 30 Hz, i.e., a total energy exposure of -25 W.mn/cm2. This exposure
condition has been determined after a few writing and testing processes in an attempt to
optimize the diffraction efficiencies.
Once the device was fabricated, we used two different methods to characterize it. The
first one consists of the measurement of the WDM device characteristics. The
collimated beam of a Ti/Sapphire tunable laser was coupled into the TE0 guided mode
with a prism and the measurements were taken from the output edge of the waveguide.
The diffraction efficiency vs. input wavelength was then recorded at each 6di. Figure 6
349
shows the results of these measurements. The diffraction efficiencies vary from 62% to
21% while the FWHMs are around 3 nm, about 6 times broader than the theoretical
values given by (2), and the cross-talks are between -10 and -15 dB.
The broader than expected FWHMs prompted us to perform a second kind of
experiment on the fabricated samples to inspect the uniformity of the index modulation
profiles. By illuminating the sample with a coherent visible beam from the top (non-
guided light), we generated scattered beams, corresponding to the different orders of the
Raman-Nath diffraction through the thin layer of the grating (6 fi.m guiding layer
thickness). Because the intensity of each of them is proportional to the local index
modulation of the grating the use of a large diameter incident beam allows the
observation of the index modulation profile over the whole grating area. As suspected,
highly non-uniform gratings have been observed in the samples fabricated. We attribute
~
Ge (7%): Si02 5mm
Thick.: 6 urn
An: 0.7
SiOz
Thick.: 20 urn
«
•3 +32ü
o 6dl ==
c 0.6
■S» %2--= -26°
o
ed3 == -30°
c em == +28°
.o
Ü
I 0.4
0.2
0.0 ==■
820 830 840 850 860 870
Input wavelength (nm)
Figure 6: Measured diffraction spectrum of each port of the 4-channel WDM.
The inset shows the guiding structure and the WDM geometry.
350
this to the non-uniformity of the excimer laser beam itself. Indeed, the particular laser
we have in our facility was working at just above the threshold because a Fabry-Perot
etalon was installed inside the cavity to obtain the minimal coherence length needed for
the holographic writing.
The non-uniformity of the index modulation over the gratings area explains the
broadening of the spectral response and also the relatively high level of cross-talk. It
also partially explains the relatively low diffraction efficiencies, which should be
improved with a better control of the gratings strength, i.e., of the exposure.
However, as a proof-of-concept demonstration, the 4-channel WDM implemented in
silica glass exhibited satisfactory functionality of wavelength demultiplexing.
The feasibility of integration of passive and active optical devices in glass has been
demonstrated. Future projects will extend the integration.
A realistic goal for the near future on the laser front is an amplifier in an optical
cavity, formed by a distributed Bragg grating written onto the guide and including
directional couplers, wavelength multiplexors, optical fibre connectors and isolators.
Another possibility is a system of 3 dB splitters making a 1 x 8 fan-out with zero
insertion loss. Switching, using pump induced refractive index changes has been
demonstrated elsewhere [18].
On the WDM front, the concept of supergrating WDM in a planar waveguide, built
in special glasses, photorefractive materials or semiconductors and the device design
method can be further extended to various new configurations and implementation
schemes. Further improvements to the device performance are shown to be possible
through both theoretical analysis and experiments. Implementations in other material
systems, including the Er:Ytterbium doped phosphate glass, are being carried out with
preliminary success.
5. Acknowledgements
6. References
1. G. H. Chartier, P. Jaussaud, A. D. de Oliveira, O.PanimK, Electron. Lett., 1977, Vol. 13,
pp. 763.
la. H. Aoki, O. Maruyama, Y. Asahara, "Glass waveguide laser",IEEE Phot. Tech. Utters,
1990, Vol. 2, No. 7, pp. 459-460.
2. D. Barbier, J. Hubner, J.M. Jouanno, A. Kevorkian, A. Lupascu, B.Hyde, "Waveguide
amplifiers in rare-earth doped glasses: fabrication, characterisation, modelling and
351
S. A. GUREVICH
A. F. Ioffe Physico-Technical Institute, RAS
26, Politekhnicheskaja, St Petersburg, 194021, Russia
1. Introduction
During last two decades the transmission capacity of fiber-optical communication sys-
tems has been increased for about one order of magnitude in each five years. Now we
are at the edge of 10 Gbit/s. Further increase in transmission rates, up to 40 Gbit/s, is
considered as a goal to be reached at the end of this century. The main purpose of such
systems is to establish local and global computer networks capable to operate with
huge amount of information. This is, of course, rather challenging engineering prob-
lem.
Two general approaches are useful in realization of extremely high transmission
rates in fiber-optical communication links. They are time division multiplexing (TDM)
and wavelength division multiplexing (WDM). TDM implies very broad operation
bandwidth of diode lasers, photodetectors and driving electronic components, the inte-
grated optoelectronic transmitter and receiver to be the only practical design in this
case. In WDM systems, the precise control of multiple wavelength in operation is of
grate demand. Regarding diode lasers as a major light source for TDM and WDM
systems, the question is whether these lasers can meet the specific requirements im-
posed by high operation frequencies and desired high degree of wavelength control. In
diode lasers directly modulated by pumping current the modulation bandwidth is lim-
ited by several factors like differential and nonlinear gains, carrier drift time across the
waveguide, carrier capture time into the active layer, the maximum reported bandwidth
being 33 GHz [1]. Characteristic for directly modulated diode lasers is also the dy-
namic chirping of the emission wavelength [2]. At transmission rates above 10 Gbit/s
chirp may create severe problems if the length of line is about 100 km. Besides, there is
temperature related drift of emission line (about 1Ä/K), which prevents from precise
wavelength control. This is especially worth effect when WDM is used.
However, several new ideas are circulating which may be useful in designing of
diode lasers free from the above discussed shortages. Practical realization of these
propositions might be a breakthrough for extremely high transmission rates.
353
5. Luryi et al. (eds.), Future Trends in Microelectronics, 353-364.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
354
Recently, dual modulation technique has been proposed [3], which implies that
pumping current is modulated simultaneously with one of the parameters controlling
the optical field in the laser cavity. Possible candidates are material gain, optical con-
finement factor or mirror reflectivity. With appropriate choice of amplitudes and
waveforms of two laser controlling signals, the frequency dependent laser output re-
sponse may be substantially flat, promising the modulation bandwidth up to 100 GHz.
Dual modulation results also in radical suppression of chirp by rejecting carrier con-
centration oscillations inside the laser active layer. Below, we will examine in some
detail the advantages of laser operation under dual modulation of the pumping current
(J) and optical confinement factor (T). In the consideration of dual J&r modulation
scheme we refer to the specific laser structure proposed earlier for confinement factor
control [4]. Besides, we will look at new design of the diode laser which allows for
efficient control of the cavity losses via governing the laser mirror reflectivity. In some
cases, dual modulation by pumping current and mirror reflectivity, J&R modulation,
may turn to be even more promising for high speed operation when compared to J&r
modulation.
The spectral position of the laser emission line may, in general, be stabilized by
using extended cavity lasers, where the gain area is only a small part of the hole reso-
nator length. In this paper we are considering new laser design based on integrated
optic technology, which allowed earlier to effectively match on a common substrate the
heterostructure laser waveguide with dielectric-film waveguide [5]. In composite cavity
laser, having integrated distributed Bragg reflector on dielectric-film waveguide and
extremely short gain region, both the emission line drift and chirp are supposed to be
reduced by two orders of magnitude compared with common diode lasers.
2. Possible High Speed and Low Chirp Operation of a Diode Laser Modulated by
Dual Pumping Current and Optical Confinement Factor Control
An important factor limiting the modulation bandwidth of a diode laser is that the in-
fluence of pumping current on the photon density is not straightforward but takes place
via the variation of nonequilibrium carrier density in the laser active layer. As a result,
even in the ideal case of instantaneous carrier drift across the waveguide, fast carrier
capture into QW active and lack of circuit parasitics, the intrinsic output modulation
response of the laser drops rather abruptly with the increase of modulation frequency.
The situation may be quite different and the modulation bandwidth can be substantially
expended if the pumping current is controlled simultaneously with the confinement
factor. Demonstration of enhanced operation frequencies and chirp suppression by dual
modulation of optical confinement factor and pumping current is the subject of on-
going research work.
355
Ipump
n^GaAs
contact
layer \^
undoped
waveguide
+ QW
Figure 1. Four-terminal ridge guide laser structure and its connection scheme.
In the following we consider the laser structure of a ridge guide type [4], having
four electrodes (Fig. 1). Single quantum well AlGaAs/GaAs separate confinement het-
erostructures were used for the device fabrication. Two grooves were etched in parallel
to define the ridge and to separate the ridge from the side areas. Two contact terminals
placed on the ridge top (5 |xm wide) and on the substrate surface were used for laser
356
pumping, whereas two electrodes situated on the top surface of the side areas were for
governing the potential distribution in the structure. The laser cavity length was nor-
mally 500 |om. Fig. 1 shows also the laser connection. The current generator, Ipump, is
connected between the ridge and the substrate. Two voltage generators, Usidei and Uside2,
are connected between the ridge and side contacts to produce additional bias of the
structure. With this, near field intensity distributions were measured, scanning the im-
age of the laser mirror in lateral direction, parallel to the junction plane. In these ex-
periments the subject of main interest was the dependence of near field patterns on the
voltage applied to the side contacts.
Fig. 2 shows normalized near field patterns recorded at different Us voltages and
pumping currents. Below the laser threshold, FWHM of near field distributions de-
pends on side voltage if it is in the range Us=l+3 V. It means that Us controls the cur-
rent spread in the structure and thus governs the lateral gain/loss profile. Above the
threshold, Us voltage was found not to influence FWHM of the near field pattern corre-
sponding to the lateral mode shape. This is because the lateral mode shape is deter-
mined by index guiding ridge. Consequently, Us voltages influences the overlap of the
gain and optical mode which is the mechanism of optical confinement factor control.
To analyze the dynamic behavior of our four-terminal laser we carried out small signal
analysis of laser rate equations. Basing on previous results, we assume that at suffi-
ciently high Us the size of inverse population area w is controlled solely by the side
voltage. In the laser with the cavity length Z, the current density J is given by:
I
J=— (1)
w-l
where I is the overall pumping current. The variation of (1) gives:
5J 51 8w
= l w
(2)
Jo o o
The optical confinement factor r should also be considered as a function of w.
Using the effective refractive index approximation, the optical confinement factor can
be represented as a product of its lateral and transverse parts:
r = rrr,. (3)
Assuming carrier density and gain are constant over the inverse population area of
width w, we have:
W -
j\E(x)\ dx
]\E(xfdx
357
where E(x) is the lateral mode field, x - axis is paralleled to the junction plane. We will
consider I and Us as driving forces, causing the variations of w, Th J and, consequently,
of carrier and photon densities. The variation of optical confinement factor can be de-
rived from (3) and (4) as:
ÖT |£(x)|28w
(5)
L
° j\E(x)\ dx
o
Defining the mean value of optical mode intensity as:
E2-w = ]\E{xfdx, (6)
o
and introducing the ratio of optical field intensity at the point x=w to its mean value:
\E(w)f
* = -fT-. <7>
we obtain:
8r 5w
r w
o o
The parameter % characterizes the efficiency of optical confinement factor modu-
lation. As it is clear from (7), £=1 when w is small compared to lateral size of the opti-
cal mode and £-»0 if w is comparable or even large with respect to mode size. The
value of E, is determined by side voltage as well as by the parameters of ridge
waveguide. In our case §=0.1 when Us = 1.0 V (see Fig. 2). Further, neglecting the
influence of parasitics in the laser structure we assume that temporal variation of w
immediately follows the variation of Us. With this, we have:
ow 8US
— = -C— • (9)
W U
0 s0
tages of this modulation regime is that relaxation oscillations are suppressed and
modulation bandwidth is very broad.
Providing Sn=0, the linearized rate equations are:
8J
vg05S=—, (10a)
ed
ß"o
8S+A5S = (vg0SQ + )Sr, (10b)
others notations are common. Remarkable feature of equation (10a) is that the varia-
tion of photon density is just proportional to the variation of current density. If the
amplitude of the current density variation SJ is kept constant by appropriate choice of
SI and ör, the output response of the laser will be frequency independent. It means that
in practice the modulation bandwidth will be limited only by the structure and circuit
parasitics.
Eliminating SS from equations (10a) and (10b), we have:
/co + A j'(co)
<u)
*«»- K^T-
ß«o
where y((o) snAj(co) are complex amplitudes of STand SJ variations, respectively, and
co is the modulation frequency. Frequency dependent relation (11) should be main-
tained in order to keep Sn=0.
We will consider the case when side voltage and pumping current are modulated
so that the relation between the amplitudes of these signals obey the following relation:
/(co) u(co) j
+C = — = const (12)
J U
0 s0 Jo
To fulfill (11), the following additional relation should be satisfied:
r0 /co+A j
w <13)
« ^r V£o(v£o5'o+—)
wn
1,
359
2.3. RESULTS OF NUMERICAL SIMULATIONS
00 ni SI/Io
a U»o= 1.5 V
a> M
0.2
° s
II 1
°n 0.0
S «=
SI " -0.1
"M ^
p,^ -0.2
o
2; -0 1
-0.4 —
-0.5
In this section, we consider the possibility of the laser output modulation by controlling
the pumping current and the mirror reflectivity. Compared to the laser structure stud-
ied for optical confinement factor modulation, new laser structure is even more tailored
for high speed operation. To the best of our knowledge, there was only one paper [6]
where the laser with tunable reflectivity mirror was described That was surface emit-
ting laser with the reflectivity of DBR controlled by electrooptic variation of the layer
refractive indexes. However this structure is rather complicated in fabrication and can
hardly be considered as practical one.
The laser structure having integrated Y-branch tunable mirror is shown schemati-
cally in Fig. 4. When the phase difference between two waves coming back after
propagation through Y-branch reflector is equal to odd integer multiplied by TC, de-
structive interference takes place in the input waveguide. In this case, the optical power
is radiating into substrate modes at the Y-point and the reflection from Y-branch sec-
tion is near to zero. Contrary to this, if the phase difference is even integer multiplied
by 7t, the reflection from ideal Y-branch mirror is equal to unity. The variation of
phase difference in Y-branch reflector and thus the modulation of its reflection coeffi-
cient can be accomplished by using the dependence of the refractive index on carrier
concentration or electric field which are sufficiently strong effects in QW containing
A3B5heterostructure materials.
cleaved mirror reflection
/ coating
To eliminate the optical losses in the tunable mirror, the waveguide in Y-branch
area can be formed by using impurity induced disordering, ion beam disordering or
selective growth techniques. Next, dielectric is deposited on the structure surface and
the windows are aligns and opened on the top of the gain and phase controlling areas
361
and contacts are formed. Reflection coating is produced at the end of Y-branch
waveguides. It is to note that integrated Y-branch reflector placed inside the laser cav-
ity can be used as tunable mirror in combination with DFB structure in the gain area.
As most of electrooptic devices, integrated Y-branch reflector can be tuned at
rather high frequencies. Being limited by only RC parameters of the circuit, the opera-
tion bandwidth of the reflector itself may be about several dozens of GHz. Note that,
the above described design insures strong electrical separation of gain and phase con-
trolling areas. Owing to this, the electrical signals governing the gain and phase con-
trolling areas will not intermix. This feature is important when dual modulation
scheme is applied. The above described structure may be superior over that designed
for optical confinement factor modulation, where the maximum speed of operation is,
in practice, more complicated function of structure parasitics. Thus, in the laser with
tunable Y-branch reflector all the advantages of dual modulation scheme can be real-
ized.
In single frequency DFB and DBR heterostructure diode lasers the stability of the
spectral position of single longitudinal mode is an important parameter. There are two
sources of emission line drift in DFB and DBR lasers. One is related to the variation of
carrier concentration occurring under hf output modulation. This is the reason for dy-
namic line chirp. The other source of line drift is associated with temperature depend-
ence of the refractive indexes of waveguiding layers. To some extent it could be elimi-
nated by introducing the electrical feedback loop to the laser mount cooler but this will
enhance complexity and cost of the laser module. Dynamic chirp in DFB and DBR
laser can be eliminated by appropriately choosing coupling coefficients and grating
phase shifts or by using external modulators. However, the chirp reduced in these ways
is still fare above the transfer limited one [7]. Lasers operating free from both excess
dynamic chirp and temperature related line drift would be very useful for high capacity
communication systems.
Our proposition for the lasing line stabilization is based on the results of previous
work [5], where we have first developed the integrated optic technology which makes it
possible efficient end-to-end matching of heterostructure laser waveguide with thin-
film dielectric waveguide sputtered on a common semiconductor substrate. Using this
technology, single frequency heterostructure lasers with monolithically integrated DBR
on corrugated dielectric waveguide have been fabricated. Because of very small tem-
perature coefficients of refractive indexes of dielectric waveguiding layers, the spectral
position of DBR reflection band was extremely stable. The width of DBR reflection
band being = 2 A, its temperature drift was as small as = 0.01 Ä/K. Owing to high
waveguide matching efficiency, low optical loss in dielectric DBR and because of high
regularity of DBR parameters, the laser output efficiency was high, 32% per cleaved
362
output mirror. Besides selection of single longitudinal mode, dielectric film DBR of-
fered strong selection of fundamental lateral mode.
However, the stabilization of DBR reflection band is necessary but not sufficient
for the emission line stabilization. Indeed, in DBR lasers discussed above the variation
of device temperature led to the variation of refractive index of heterostructure gain
area and due to this the longitudinal cavity modes drifted through narrow and position
fixed DBR reflection band. As a result, mode switching was observed under the tem-
perature variation. To avoid mode switching one should, in addition, fix the position of
the cavity mode.
This may be achieved in composite cavity DBR laser fabricated by using the above
discussed integration technique. The proposed laser structure is shown schematically in
the Fig. 5. In this laser the heterostructure gain area has relatively small length 1 and
heterostructure waveguide is monolithically and-to-end joint with dielectric waveguide.
The corrugation of the dielectric waveguide starts at the distance L from the waveguide
matching interface. Thus, DBR laser has composite cavity with the length of regular
waveguide area equal to 1+L. For the laser to operate in single frequency mode, the
length 1+L should be taken so that the corresponding intermode distance be compara-
ble to the width of DBR reflection band. On the other hand, to make the emission line
position insensitive to the variations of both temperature and carrier concentration, the
condition 1/L « 1 should held. The other important condition involved is that the light
reflection at the waveguide matching interface should be small. This condition seems
to be easily fulfilled by using previously developed integration technology.
pumping contact
heterostructure
waveguide DBR dielectric
output waveguide
<= // // // // // // // // // // // // // // // // // // // /. 1
semiconductor substrate
Figure 5. Schematic drawing of the laser with integrated dielectric film waveguide DBR.
5. Summary
In this paper we have discussed several possibilities for the diode lasers to have consid-
erably expanded modulation bandwidth, low chirp and high terminal stability of the
laser line spectral position. These are the features of main concern in the design of
future very high-bit-rate fiber-optical communication systems. It was shown that dual
modulation of the optical confinement factor and laser pumping current may result in
operation bandwidth up to 100 GHz and in very low chirp. New laser structure with
tunable Y-branch mirror is proposed for realization of dual modulation by governing
the pumping current and cavity loss, which can be even more prospective for high
speed operation. Novel structure of integrated composite cavity DBR laser is consid-
ered for suppression of dynamic chirp and thermal drift of the lasing line, promising
the improvement of these parameters by two orders of magnitude.
Acknowledgments
The author would like to thank M. Shatalov, Dr. V. I. Skopina and E. Tanklevskaya
for their essential assistants in this work, O. Utkina for the help in preparation of the
manuscript. Useful discussions with Prof. R. A. Suris are greatly acknowledged. This
work was supported in part by ISF grant NU 9000.
References
1. Ralston J.D., Eisele, K., Sah, R.E., Larkins, E.C., Weisser, S., Rosenzweig, J.,
Fleissner, J., and Bender, K. (1994) Low-Bias-Current Direct Modulation up to 33
GHz in GaAs-Based Pseudomorphic MQW Ridge-Waveguide Lasers Suitable for
Monolithic Integration, Conference Digest of 14th ISLC, Maui, USA, 211-212.
2. Bowers, J.E. (1987) High speed semiconductor laser design and performance,
Solid- State Electronics 30, 1-11.
364
3. Gorfinkel, V.B., Camacho, F., and Luryi, S. (1993) Dual Modulation of Semicon-
ductor Laser, Proceedings of 1993 International Semiconductor Device Research
Symposium (ISDRS), Charlottesville, Virginia, USA, 723-726.
4. Gorfinkel, V.B., Kompa, G., Gurevich, S.A., Shtengel, G.E., and Chebunina, I.E.
High Frequency Modulation of a QW Diode Laser by Dual Modal Gain and
Pumping Current Control, Proceedings of 20th International Symposium on GaAs
and Related Compounds, Freiburg, Germany, 41-42.
5. Alferov, Zh.L, Gurevich, S.A., Karpov, S.Yu., Portnoi EX., and Timofeev, F.N.
(1987) Monolithically-Intergrated Hybrid Heterostructure Diode Laser with Di-
electric-Film Waveguide DBR, IEEE Journal ofQuant.Electron. QE-26, 869-881.
6. Blum O., Zucker, J.E., Chiu, T.H., Divino, M.D., Jones, KL., Chu, S.N.G., and
Gustafson, T.K. (1991) InGaAs/InP multiple quantum well tunable Bragg reflec-
tor, Appl.Phys.Lett. 59,2971-2973.
7. Jonson, J.E., Tanbun-Ek, T., Chen, Y.K., Fishman, D.A., Logan, R.A., Morton,
P.A., Chu, S.N.G., Täte, A., Sergent, A.M., Sciortino, P.F., Wetch, K.W., and Jr.
(1994) Low-Chirp Integrated EA-Modulator/DFB Laser Grown by Selective-Area
MOVPE, Conference Digest of 14th ISLC, Maui, USA, 41-42.
INCREASED-FUNCTIONALITY VLSI-COMPATIBLE
DEVICES BASED ON BACKWARD-DIODE
FLOATING-BASE Si/SiGe HETEROJUNCTION
BIPOLAR TRANSISTORS
Z. S. GRIBNIKOV
Institute of Semiconductor Physics
Ukrainian Academy of Sciences
Kiev-28, Ukraine
S. LURYI
Dept. of Electrical Engineering
SUNY at Stony Brook
Stony Brook, NY, U.S.A. 11794
A. ZASLAVSKY
Div. of Engineering
Brown University
Providence, RL U.S.A. 02912
1. Introduction
2. Device operation
In this section we illustrate the principle of operation3.4 of increased-
functionality HBT devices built in silicon. Consider a Si/SiGe/Si npn
bipolar transistor with a backward diodes emitter-base (E-B) np junction
(created by the appropriate doping of the emitter and base regions) and a
standard base-collector pn junction. The emitter region is contacted by two
(or more) trench-isolated electrodes, while the base is left floating, as
shown schematically in Fig. 1. If the E-B junction is forward-biased,
minority electrons are injected into the base and extracted by the collector,
resulting in ordinary bipolar transistor operation with current gain.
However, because of the backward-diode characteristic shown in Fig. 2,
when the E-B junction is reverse-biased it acts as an efficient Ohmic
contact to the base.
If in the structure of Fig. 1 both of the emitter contacts are grounded
(Vel = VQ2 = 0) and the collector is biased high (Vc » kT), a small
collector current will flow as in a standard bipolar transistor with a floating
base. Conversely, if one of the emitter contacts is grounded (Vel = 0) and
the other is biased Ve2 < Vc, a large hole current will flow through the
reverse-biased E-B junction under the second emitter (which acts as a
"base" contact), leading to a large output current. If the current densities
of the forward and reverse-biased junctions are Jj and J2, and
corresponding ponding junction areas are Ai and A2 respectively, in the
active transistor regime with high common-emitter current gain (ß » 1)
367
emitter 1 emitter 2 4i
1 sEZÄ&zzza
backward
diode E-B
junctions
collector V
n collector
we have:
3. Conclusion
The implementation of proposed backward-diode floating-base HBT's in
Si/SiGe heterostructures is advantageous from both the processing and
VLSI integration standpoints. Since contacting the base is typically the
most demanding task in bipolar transistor fabrication, the proposed
structures would greatly reduce the number of processing steps. The
integration of these bipolar devices with standard CMOS designs will
benefit from the fabrication simplicity, leading to easy implementation of
such promising BiCMOS circuits as CMOS logic integrated with high
current-drive bipolar Si/SiGe input-output devices.
(a)
(b)
FIG. 4. Band diagrams of the device under the two emitter contacts under
the Vel = high (a — injecting emitter) and Ve2 = low (b — contact
emitter). The base is taken to be Sii.xGex with the Ge fraction x = 0.2
(graded up from x = 0 in the low-doped base-collector junction).
370
4. References
1
Hess, S., Morkoc, H., Shichijo, R, and Streetman, B. G. (1979)
Appl. Phvs. Lett. 35, 649; Luryi, S., Mensz, P. M., Pinto, M. R.,
Garbinski, P. A., Cho, A. Y., and Sivco, D. L. (1990) Appl. Phvs.
Lett. 57, 1787.
2 Sen, S., Capasso, F., Cho, A. Y., and Sivco, D. L. (1987) IEEE
Trans. Electron. Dev. ED-34, 2185; Seabaugh, A. C, Kao, Y.-C,
and Yuan, H.-T. (1992) IEEE Electron Dev. Lett, 13, 479.
3
Imamura, K., Takatsu, M., Mori, T., Bamba, Y., Muto, S., and
Yokohama, N. (1994) Electron. Lett. 30, 459; Imamura, K. etai,
Extended Abstracts 1994ICSSDM, Yokohama, pp. 467-469.
4
Gribnikov, Z. S. and Luryi, S. (1994) Article comprising a bipolar
transistor with a floating base, AT&T Bell Laboratories patent item
1-32, filed August, 1994.
5
Sze, S. M. (1981) Physics of Semiconductor Devices, 2nd ed., Wiley,
New York, p. 537.
6
For a review of Si/SiGe heterostructures see, for example, Bean, J. C.
(1992) Proc. IEEE SO, 571.
7
Robbins, D. J., Canham, L. T., Barnett, S. J., Pitt, A. D., and
Calcott, P. (1992) J. Appl. Phys. 71, 1407.
REAL-SPACE-TRANSFER OF ELECTRONS
IN THE INGAAS/INALAS SYSTEM
W. TED MASSELINK
Humboldt-Universität zu Berlin,
Institut für Physik,
10099 Berlin, Germany
1. Introduction
Monte-Carlo analysis [1] indicates that as FET gate lengths shrink to the
sub-100 nm regime, the electric fields in a field-effect transistor become
large enough that the electrons occupy most of the Brillouin zone. Because
carriers outside of the T valley behave similarly in most semiconductors,
transistor performance in this small-gate-length limit is predicted to be
rather material-independent.
A noted exception to this generalization is 1no.53GüQ,47As (i.e., InGaAs
lattice-matched to InP), which is predicted to significantly outperform most
other semiconductors. The advantage in InGaAs is that the energy differ-
ence between the lowest-lying conduction-band T valley and the higher-
lying L and X valleys is much greater than that in most other semicon-
ductors. Thus the electrons in InGaAs tend to remain in the V valley to
larger electric fields, resulting in a higher peak velocity and therefore in
a higher average velocity with correspondingly smaller transit time in the
transistors.
Ref. [1] assumes that each semiconductor is used in a MOSFET struc-
ture, in which the channel is implanted p-type and the gate is insulated
with S1O2, whose conduction band lies much higher than the conduction
band of the semiconductor comprising the channel. InGaAs FETs, however,
are usually of a heterostructure design, in which the electrons are con-
fined by the wider-bandgap Ino.^Alo.^As. Because the conduction band
of the 7710.52^0.48^ ues only 0.5 eV above the conduction band of the
Ino.53Gao.47As [2], while the L valleys of the Ino.53Gao.47As lie 0.55 eV
above the T valley [3], scattering of channel electrons out of the channel into
371
S. Luryi et al. (eds.), Future Trends in Microelectronics, 371-376.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
372
the InAlAs gate insulator (real-space transfer [4]) may be more important
than the intervalley scattering. The present experimental results demon-
strate that the peak electron velocity in Jno.53Ga0.47^W^o.52^0.48^-s
modulation-doped heterostructures is smaller than in bulk Ino.53GaOA7As
and indicate that the additional scattering mechanism responsible for this
behavior is real-space transfer.
The use of pseudomorphic InGaAs with enhanced In content can im-
prove the transport characteristics of InGaAs-containing heterostructures
by decreasing the electron mass and increasing the energy separation be-
tween the InGaAs T valley and the InAlAs T valley. Further results of mate-
rial transport and FET characteristics in pseudomorphic structures indicate
that strain can be used to increase the bandgap compared to unstrained
material, thus allowing higher breakdown voltages, while simultaneously
improving the electron mobility and peak velocity.
30 —■ 1 ■ r
16 -3
n=3x10 cm
300 K ln
E
o
o.53Gao.47As
(O
o
2 4 6 8
Electric Held (kV/cm)
30
- 300 K
ln
0.53Ga0.47As
InGaAs/lnAIAs
2 4 6 8
Electric Field (kV/cm)
which to scatter in the InAlAs, we expect these two effects to account for
the peak velocity in the heterostructure occurring at a lower electric field
and therefore smaller magnitude when compared to bulk. Simple numerical
calculations similar to those of McCumber and Chynoweth [10] also support
the hypothesis that the added scattering channel from the two-dimensional
electron gas (2DEG) into the InAlAs can account for the reduced peak
velocity [11].
One implication of this result is that in order to attain the ultimate
performance potential in InGaAs-based FETs, a wider-bandgap insulator
in place of the IUQ^AIQ^AS would be advantageous to reduce real-space
transfer. Because of the difficulty in depositing a wide-bandgap insulator
such as SiÖ2 onto the InGaAs and simultaneously maintaining a defect-free
interface, the best compromise may be to use pseudomorphic InyAli-yAs
with excess Al (y < 0.52), thereby increasing the conduction band discon-
tinuity [12].
Alternately, or in combination with the use of a wider-bandgap insu-
lator, the structure can benefit from the use of pseudomorphic InGaAs
with excess In. This material has a lower-lying T valley, resulting in an
increased T-L energy separation. In related work [8], we have optimized
one such variation of this idea with a strained InAs quantum well in the
middle of a (lattice-matched) InGaAs well. Unstrained InAs has a bandgap
of only 0.36 eV with L valleys lying 1.08 eV above the T valley. When the
InAs is pseudomophically grown on the InP lattice constant, the bandgap
will be approximately 0.55 eV. Further, in a 30-Äwell width, the confine-
ment effects will effectively increase the InAs bandgap to approximately
0.65 eV. The InAs first electron subband is still lower than the T valley
in the InGaAs so that electron transfer (either k-space into the L valleys
or real-space into the adjacent InAlAs) is reduced through the presence of
the InAs. Additionally, the relatively wide effective bandgap of the InAs
quantum well ensures that under high electric fields, impact ionization will
not be a limiting scattering mechanism.
3. Conclusion
References
1. M.V. Fischetti and S.E. Laux, IEEE Trans. Electron Devices 38, 650 (1991).
2. R. People, K.W. Wecht, K. Alavi, and A.Y. Cho, Appl. Phys. Lett. 43, 118 (1983).
3. K.T. Cheng, A.T. Cho, S.B. Christman, T.P. Persall, and J.E. Rowe, Appl. Phys.
Lett. 40, 423 (1982).
4. K. Hess, H. Morkog, H. Shichijo, and B.G. Streetman, Appl. Phys. Lett. 35, 469
(1979).
5. W.T. Masselink, Semicond. Sei. Technol. 4, 503 (1989).
6. D. Hahn and A. Schlachetzki, J. Electronic Mat. 21, 1147 (1992).
7. M.V. Fischetti, IEEE Trans. Electron Devices 38, 634 (1991).
8. J.K. Zahurak, A.A. Iliadis, S.A. Rishton, and W.T. Masselink, J. Appl. Phys. 76,
7642 (1994); J.K. Zahurak, A.A. Iliadis, S.A. Rishton, and W.T. Masselink, IEEE
Elec. Device Lett. 15, 489 (1994).
9. W.T. Masselink, N. Braslau, W.I. Wang, and S.L. Wright, Appl. Phys. Lett. 51, 1533
(1987).
10. D.E. McCumber and A.G. Chynoweth, IEEE Trans. Electron Devices 13, 4 (1966).
11. W.T. Masselink, Appl. Phys. Lett. 67, (to be published, 1995).
12. S.R. Bahl, W.J. Azzam, and J.A. del Alamo, IEEE Trans. Electron Devices 38,
1986 (1991).
CHARGE INJECTION TRANSISTOR AND LOGIC ELEMENTS IN
Si/Sii_xGex HETEROSTRUCTURES
M. MASTRAPASQUA
Eindhoven University of Technology
600 MB Eindhoven, The Netherlands
1. Introduction
The charge injection transistor (CHINT) [1] concept refers to a class of devices based on
the principle of real space transfer (RST) [2] of hot carriers between two independently
contacted conducting layers. One of these layers, the emitter, has source and drain
contacts, while the other, collector, layer is separated by a heterostructure barrier. When
an electric field is induced between the drain and the source, the carriers in the channel
become "hot" and can overcome the barrier. The RST effect manifests itself in two
ways. The collector current Ic, at a constant collector bias Vc, increases at a
sufficiently high source and drain bias. Simultaneously, the drain current decreases
showing a negative differential resistance (NDR) in the current voltage characteristic.
A fundamental property of RST transistor is that the collector current does not
change if the source and drain contacts are interchanged. Thus the device exhibits an
exclusive-OR (xor) dependence of the output current on the input voltages regarded as
binary logic signals. Even more powerful logic functionality is obtained in a CHINT
device with three input terminals [3]. This device, which we shall refer to as the
ORNAND gate, has a cyclic three-fold symmetry. Depending on the logic value, high
or low, of one of the three electrodes, the output current behaves as an nand or or
function of the other two electrodes.
The charge injection transistor has been successfully implemented in a number of
III-V systems [1, 4-6]. A monolithic optoelectronic ORNAND device has also been
demonstrated in a InGaAs/InAlAs heterostructure material [7]. However, there is great
interest in realizing the charge injection transistor in a silicon-based heterostructure,
making it possible to integrate CHINT devices with traditional VLSI silicon logic and
memory circuits. Hot hole RST has been observed in Si/SiGe heterostructures by
Mensz et al. [8]. Recently, RST of electrons in a strained silicon layer has been
observed in an n-type Si/SiGe heterostructure [9]. The present work reviews our recent
results on the realization of CHINT, and a monolithic ORNAND function logic device
with an epitaxial layer structure containing strained SiojGeoj and Si [10].
377
S. Luryi et al. (eds.), Future Trends in Microelectronics, 377-383.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
378
2. Epitaxial Structure and Device Fabrication
Figures 1 and 2 show schematic cross-sections of the device structures discussed below:
the CHINT is illustrated in Fig. 1 and the ORNAND gate in Fig. 2. We grew the
Si/Sii_xGex by rapid thermal epitaxy (RTE) on a 125 mm p-type Si substrate. The Ge
fraction (x) in the emitter channel and in the collector layer was 0.3, as determined by
Rutherford backscattering spectrometry. The strained layer thickness was about 15 nm,
as measured by transmission electron microscopy (TEM) and secondary ion mass
spectrometry (SIMS), which is less than the mechanical equilibrium critical thickness
for this composition.
r^tsaiPi
SOnm.SI
\ i
cap layar (undopad)
V
15 nm, Sl07ee0 j channel (undopad]
c
15nm,si07Oe03collector (B:3xltfB)
1000 nm, SI
SI
buftar layar (B: 5x10*8)
substrata (B:2x1o")
W
Figure 1. Schematic cross section of the charge injection transistor structure. The
distance between source and drain, Lch = 0.5 |im and the width W =40 urn.
UndopedStj.7Gea3Ctanml
'UndopedSICap
palate Collector
C
Figure 2. ORNAND logic gate structure. Each of the three channels, defined by the
p+ diffusion in the silicon cap layer, has length Lch =1.0 |J.m and width W = 40 |i.m.
Cyclic symmetry, required for the ornand logic operation, results from
the periodic boundary condition on V3.
379
Shallow source and drain ohmic contacts are the most critical processing step for
CHINT devices, because the contact must be deep enough to penetrate through the cap
layer into the 15 nm channel without reaching the underlying barrier layer. We used
RTE to grow a B-doped Ge layer to act as a diffusion source [11]. The Ge layer grows
selectively only on the exposed silicon surface and not on the oxide. After growth, we
used rapid thermal annealing for 12 min. at 800 °C to form the junction contacts with
an estimated final depth < 90 nm and active surface concentration >1020 cm"3.
Finally 10 nm Ti, 100 nm TiN and 500 nm of aluminum were deposited on the
front of the wafer and patterned to form the source, drain and collector contacts. Since
most of the band gap discontinuity between the strained Sirj.7Geo.3 channel layer and
the Si barrier falls into the valence band, the devices employ RST of hot holes.
3. Device Characteristics
-2.5
-1 -2 -3 -1 -2 -3
Figure 3. Room temperature characteristics of the drain (a) and collector (b) current
at different collector biases for a 0.5 |lm x 40 ^m Si/SiGe CHINT.
380
4. Logic Functions
Figure 4 demonstrates the xor logic operation at two different lattice temperatures. We
see that IQ = xor(Vn,Vs) at all temperatures. The on/off ratio is 10 dB at room
temperature and increases to 65 dB at cryogenic temperature due to diminishing leakage
current in the (0,0) logic state.
time
o
1 ■
-1 •
lc TL = 295 K
(mA) -0.12
-0.8»
lc TL = 77 K
(mA) 0.18 nA;
V3 time
0
1*
V2 0
Vi o
AL,'jy^
-1 .
lc
(mA)
•-0.07.
NAND OR
5. Simulations
In order to study how the device performance of the Si/SiGe CHINT can be improved,
we have used the device simulator PADRE [12] which solves Poisson, drift-diffusion
and energy balance equations for an arbitrary heterostructure device in 1D/2D/3D spatial
dimensions.
First, we have simulated the device with a structure as shown in Fig. 1, with the
doping value and layer thickness noted above. Figure 6 compares the measured electric
characteristic (dashed line) with the one simulated by PADRE (dotted line). It is
important to stress that in these simulations we have not adjusted any parameters to fit
the experimental data. Clearly, a large part of the discrepancy from simulations and
measurements may simply be due to a parasitic resistance at the source, drain and
collector contacts.
Figure 6 also shows how the drain and collector characteristic of a SiGe CHINT
would change by changing some of the parameters in the design of the structure:
-1 -2 -3 -4 0 -1 -2 -3 -4
Drain Bias VQ (V) Drain Bias VD (V)
Figure 6. Drain (a) and collector current (b) as it is simulated by device simulator
PADRE for different design parameters of the SiGe CHINT structure with
Lcj, =0.5 p.m. The dotted line is the simulation for the device structure as
shown in Fig. 1. The dashed line is the experimental data
from Fig. 2, Vc = - 4 V.
382
i) Increasing the doping level in the channel is a detriment. The channel mobility
decreases due to impurity scattering. This results in a decrease of the carrier temperature
and therefore a decrease of RST current.
ii) The absence of a silicon cap layer improves the device performance. The silicon cap
layer constitutes an alternative current path for the drain current, especially at high drain
bias. However, the use of a silicon cap layer is desirable for other reasons, although it
could be thinner than 80 nm. From a point of view of the device fabrication, the cap
layer helps to relax the requirements on the source/drain diffusion-contacts from being
abrupt and shallow to only abrupt. From the crystal growth point of view, the presence
of the cap layer promotes stability of the strained structure.
iii) A thinner barrier layer increases the overall device performance due to the increase of
the electric field over the barrier.
6. Conclusions
7. Acknowledgments
The authors would like to thank Ray Cirelli for his help in device processing and Serge
Luryi for helpful discussions.
8. References
1. Luryi, S., Kastalsky, A., Gossard, A.C., and Hendel, R.H. (1984) Charge
injection transistor based on real space hot-electron transfer, IEEE Trans.
Electron Dev. 31, 832-839.
2. For a review of real-space transfer phenomena see Gribnikov, Z.S., Hess, K., and
Kosinovsky, G.A (1995) Nonlocal and nonlinear transport in semiconductors:
real-space transfer effects, J. Appl. Phys. 77, 1337-1373.
3. Luryi, S., Mensz, P., Pinto, M.R., Garbinski, P.A., Cho, A.Y., and
Sivco, D.L. (1990) Charge injection logic, Appl. Phys. Lett. 57, 1787-1789.
383
4. Mastrapasqua, M., Luryi, S., Capasso, F., Hutchinson, A.L., Sivco, D.L.,
and Cho, A.Y. (1993) Light-emitting transistor based on real-space transfer:
electrical and optical properties, IEEE Trans. Electron Dev. 40, 250-258.
5. Belenky, G.L., Garbinski, P.A., Smith, P.R., Luryi, S., Cho, A.Y., Hamm, R.,
and Sivco, D.L. (1993) Microwave studies of self-aligned top collector charge
injection transistor, IEDM Tech. Dig., 423-426.
6. Lai, J.T. and Lee, J.Y. (1994) Enhancement of electron transfer and negative
differential resistance in GaAs-based real-space transfer devices by using strained
InGaAs channel layers, Appl. Phys. Lett. 76, 1965-1967.
7. Mastrapasqua, M., Luryi, S., Belenky, G.L., Garbinski, P.A., Cho, A.Y., and
Sivco, D.L. (1993) Multi-terminal light emitting logic device electrically
reprogrammable between OR and NAND functions, IEEE Trans. Electron
Dev. 40, 1371-1377.
8. Mensz, P.M., Luryi, S., Bean, J.C., and Buescher, C.J. (1990) Evidence for
real-space transfer of hot holes in strained GeSi/Si heterostructures,
Appl. Phys. Lett. 56, 2663-2665.
9. Zhou, G.L., Huang, F.Y., Fan, F.Z., Lin, M.E., and Morkoc, H. (1994)
Observation of negative differential-resistance in strained n-type Si/SiGe
MODFETs, Solid-State Electronics 37, 1687-1689.
10. Mastrapasqua, M., King, CA., Smith, P.R., and Pinto, M.R. (1994)
Charge injection transistors and logic elements in Si/Si i-xGex heterostructures,
IEDM Tech. Dig., 385-388.
11. Park B.G., King, CA., Eaglesham, D.J., Sorsch, T.W., Weir, B., Luftman, H.S.,
and Bokor, J. (1993) Ultrashallow p+-n junctions formed by diffusion from an
RTCVD-deposited B:Ge layer, Proc. ofSPIE 2091, 122-131.
12. Pinto, M.R. (1991) Simulation of ULSI device effects, in Cellar, G. and
Andrews, J. (eds.), Electrochemical Society Proceedings 91-11, 43-51.
New Ideology of All-Optical Microwave Systems Based on the Use of
Semiconductor Laser as a Down-Converter.
Abstract
We propose a novel all-optical structure of phase lock loop for locking two semiconductor
lasers with a stable microwave offset for use in phased-array antenna systems.
1. Motivation
OEPLL AOPLL
Laser Q Laser M
te-
Master Master """"F*
<"A
Photo
LPF
a Detector Q,;
''S
Laser
Slave i Laser
Slave
Correcting
Circuit
Local
Oscillator
CO,
Mixer
t
Correcting
Circuit
A
(Ö1F=C04-<ÖLO Q
LPF
COrF=(ÖÄ-CÖLO
2. Laser Mixer
Recently, we proposed and experimentally demonstrated [2] a novel type of
optoelectronic mixer, the laser-mixer. In this device the microwave signal is encoded in
the form of an intensity modulation of the laser optical output. The input signals to be
mixed can themselves be either lightwaves modulated in amplitude at mm wave
frequencies or conventional microwaves (or both). The essential physics underlying the
concept of the laser-mixer is external modulation of "material" parameters of the laser
controlling the propagation of light in the laser cavity, such as the modal gain or
optical losses. Consider the rate equation for photon density in the laser cavity:
'■l OSS )
at
where g is the material gain, T is the confinement factor, aioss is the cavity loss.
Modulating two quantities, say, S and a/0JJ
5' = S'0+S'1sin<»/
a =a0 +a1smG)2t
we obtain an optical response modulated at the sum and difference frequencies,
(O =al ±(02.
387
Functionally, this physics is entirely analogous to that in conventional electronic
mixers, where mixing is accomplished due to the variation of material parameters,
controlling the electric output of the device (capacitance, resistance, etc.). The idea of a
parametric control of laser output, combined with a pumping current variation, is
relatively new [3,4]. Experimentally, an efficient parametric control was demonstrated
in a three-terminal quantum-well laser structure, realized in two material systems:
GaAs/AlGaAs at the Ioffe Institute and InGaAs/InGaP/GaAs at Bell Laboratories[5,6].
Side contact Pumping
This three-terminal structure (Fig. 2) lends itself to a natural use as a laser-mixer with
electric inputs. The electric impedance of the parametric input circuit, controlling the
modal gain of this laser structure is 50 ohm, which is very attractive for microwave
applications. Another experimentally demonstrated parametric laser-mixer [2] was
based on the variation of cavity loss in a laser with saturable absorber (Fig. 3).
Double Quantum
Well
Optical Mode
Because of the short (picosecond) recovery time available in saturable absorbers, this
type of laser-mixer can take as input optical signals modulated in intensity at
frequencies well over 100 GHz. Moreover, the optical nature of the input signal closes
the entire loop in the optical medium and permits the implementation of all-optical
circuits, operating entirely with microwave signals coded as the envelope of an optical
carrier.
388
It should be clearly understood that parametric mixing in semiconductor laser is
different in principle from the heterodyne coherent mixing of optical waves. The latter
technique is based on the square-law dependence of the optical transition rate on the
lightwave amplitude | A |. In contrast, with respect to the lightwave intensity! A |2 the
semiconductor laser is a highly linear element. Mixing of different channels in an
optical communication system - which manifests itself as an unwelcome
intermodulation distortion - is an exceedingly minor effect (typically less than -70 dBc)
which arises mostly due to higher-order parametric phenomena [7].
Let us return to the AOPLL structure shown in Fig. 1 and discuss its operation
principle. Part of the radiation from both the master and the slave lasers is focused on
the fast saturable absorber, modulating its transparency at the difference frequency
CDA=QM-f2s
and phase fa. This frequency is mixed with that of mode-locked pulsation in the laser
cavity at the the latter frequency ©LO can be as high as hundreds of GHz [2,3]. It may
be further stabilized by an electronic generator. The intermediate frequency
CDlF= (DA- COLO
is then transmitted to the correcting circuit as an envelope of the optical output of the
laser mixer. The inherited phase of the intermediate signal equals «J>IF =<j>A-<ko- This
signal is received by a fast photodiode integrated in the correcting circuit, which tunes
the slave laser so as to keep fa constant.
As shown in Fig.l, the electric part of the PLL the correcting circuit is localized in the
vicinity of the slave laser. The entire optical circuit can be integrated on a single silicon
substrate. At the same time the AOPLL architecture permits a relatively remote
positioning of the master and slave lasers, which is advantageous for their stability. We
would like to emphasize the universality of the all-optical PLL architecture, rooted in
the fundamentally dispersionless transmission of the microwave envelope via optical
waveguides. Not only this implies the broad loop bandwidth but it also means that
changing the operating microwave frequency does not entail any major revision of the
passive optical circuit.
4. Conclusion
the n-th AOPLL cascade plays the role of master for the following (n+l)-st cascade.
A particular example of a system could be an array of optically coherent lasers
operating at the same wavelength. In this case, an optical signal emitted by a single
laser-master would be split between a number of AOPLL's which control the laser-
slave array, thus keeping a given unique offset frequency between the laser-master and
each of the laser-slaves.
References
1. Seeds, A.J. (1993) "Optical Technologies for Phased Array Antennas", IEICE Trans.
Electron.E16-C, 198-206.
2. Portnoi E.L., Gorfinkel, V.B., Avrutin, E.A., Thayne, I., Barrow, D.A., Marsh,
J.H. and Luryi, S (1995) "Optoelectronic Microwave-Range Frequency Mixing in
Semiconductor Lasers", IEEEJ. Select. Top. Quant. Electron., 1,451-460.
3. Gorfinkel, V.B. and Luryi, S. (1994), "Dual modulation of semiconductor lasers",
Physics and Simulation of Optoelectronic Devices, ed. by M. Osinski, Proc. SPIE
2146, 204-209.
4. Gorfinkel, V.B. and Luryi, S. (1994), "Article that comprises a semiconductor laser,
and method of operating the article", US Pat. 5,311,526.
5. Gorfinkel, V.B., Kompa, G., Novotny, M., Gurevich, S., Shtengel, G.,Chebunina, I.
(1993), "High-frequency modulation of a QW diode laser by dual modal gain and
pumping current control" 1993-IEDM Tech. Digest, 933-937.
6. Frommer,A., Luryi, S., Nichols, D.T., Lopata, J. and Hobson, W.S. (1995) "Direct
modulation and confinement factor modulation of semiconductor lasers", Appl.
Phys. Lett., 67 (Sept 18,).
7. Gorfinkel, V.B. and Luryi, S. (1995), "Fundamental limits for linearity of CATV
lasers", IEEEJ. Lightwave Techno!., 252-260.
MICROTECHNOLOGY - THERMAL PROBLEMS IN MICROMACHINES,
ULSI & MICROSENSORS DESIGN1.
Andrzej NAPIERALSKI
Division of Microelectronics & Computer Sciences
Institute of Electronics, Technical University of Lodz
Stefanowskiego 18/22, 90-924 Lodz, POLAND
e-mail: napier@j-23.p.lodz.pl
1. Introduction
The main object of this paper is to present the new problem in the modern
microtechnology which is the high power dissipation density caused by the smaller
dimensions of circuits and higher operating speeds. In the near future the thermal
problems will be the bottleneck for integration in microelectronics. Actually in the
modern silicon devices the thermal transient phenomena are as fast as electrical and
the new methods for temperature computation must be developed. The transient heat
equation can be presented in the following form:
AV2T = C,^ (1)
dt
where: T - temperature, r-time, X - thermal conductivity,
Cv - the thermal capacity per unit volume (J/m3K).
After introduction of scaling factor a, this equation can be presented under the
following form:
ä[^!LT+^!LT+_^LT]=CV_1T_ (2)
d{ax)2 d(ay)2 d{azf da21
where: x, y, z - co-ordinates
'This work was supported by European Community Basic Research Project ESPRIT (no. 8173
BARMINT), and European Community Project COPERNICUS (no. CP-940922 - THERMINIC)
391
S. Luryi et al. (eds.), Future Trends in Microelectronics, 391-396.
© 1996 Kluwer Academic Publishers. Printed in the Netherlands.
392
Therefore, if the dimensions are reduced by a factor a, the time scale has to be
reduced by a factor a2. If for example heat needs 1 second to spread over a 1 cm3
silicon block, all transistors of lfimfa = 10"4) can be heated in 10"8 second (10ns).
This is comparable to the actual electrical time constants in electronic VLSI circuits. In
the case of 0.1 (j.m technology, the thermal time constants will be as small as lOOps
and it will be absolutely necessary to consider thermal phenomena as equally fast as
the electrical!
As it can be seen there are plenty of mutual dependencies that require iterative
solving and communication between different (thermal and mechanical) parts of the
software. The exact modelling of this kind of micromachine is very complicated, and
generally not possible by an analytical solution [6]. The CAD program has to take into
account all thermal and mechanical properties of the designed structure.
393
IS
S0L
So
WmMj Bushing Ground Plan«
Fig. 2. A scanning electron micrograph of an IC- Fig. 3. a)Cross-section schematic of the micro-
processed electrostatic micromotor, fabricated motor from Fig. 2 [2];
at MIT [2] (rotor diameter = 100 \xm) b) details of acting fields and forces.
There are three possible sources of heat generation related to the micromotor
movement:
1. side-pull friction , 2. bushing friction, and 3. viscous drag. The heat generation
can be calculated assuming that all mechanical losses are converted into the heat.
One of the important question in the field of VLSI systems and power electronics is:
How to perform the thermal monitoring of the silicon wafer, containing semiconductor
devices, in order to indicate the overheating situations?
First method consists of placement of many sensors everywhere on the chip, then their
output can be read simultaneously and compared with the reference voltage recognised
as the overheating level (Fig.4).
The idea of the next method, which is proposed in this paper, is to measure the
temperature gradient along the given distance, in a few places only on the monitored
wafer (Fig.5) and evaluate obtained information in order to achieve the temperature of
the heat source [8]. This problem is known in the literature, especially in the field of
modelling of the temperature distribution, as an inverse problem. It means that we know
the results of investigated phenomena (e.g. values of the temperature in some places on
the wafer's surface) and we try to find the parameters of the phenomena's source (e.g.
the temperature of the heat sources). In general case the solution of the inverse problem
requires powerful computations and not always gives the proper result. The reason is
that a small inaccuracy in the initial conditions can cause unacceptable inaccuracy in
the final result. In case of an IC or a power semiconductor device, there is no place on
the layout for the complicated unit performing such computations, but there is also no
need for it, as we want only detect the overheating situations. Moreover, in most cases,
the overheating occurs only in one place.
: thermal JZ?
1
control r
monitored layout
-~s
tan a «■
sensot ceH 1
Fig.6 The idea of the sensor cell Fig.7 The proposed arrangement of two sensor cells .
In this paper we will consider this assumption, however the function describing the real temperature
distribution is not linear. For the given substrate one can find very precisely this function by application of the
PYRTHERM software [9]. During calculations of the heat source temperature the found function can be
transformed to the linear one.
396
The presented method can be very useful in the case of indication the overheating
situations on the surface of an IC or other semiconductor devices. It can be
implemented directly on the monitored surface without disturbing the integrity of the
device and requires only a few sensor cells placed in any way out of the monitored
area.
5. Conclusion
The main object of this paper was the presentation of the new problem in the
modern microtechnology which is the high power dissipation density caused by the
smaller dimensions of circuits and higher operating speeds. Actually in the modern
silicon devices the thermal transient phenomena are as fast as electrical and the new
methods for temperature computation must be developed. In this paper some thermal
problems related to modern microtechnology have been discussed.
The three examples of the modern devices based on the thermal phenomenon
have been presented. As the conclusion, one can say, that in the next future much more
modem devices based on the thermal phenomenon will be certainly designed.
6. References
1. Bart, S.F. (1990) Modeling and Design of Electroquasistatic Microactuators, Ph.D. dissertation,
Massachusetts Institute of Technology, Cambridge,
2. Bart, S.F., Mehregany, M., Tavrow, L.S., Lang, J.H., Senturia, S.D. (1992) Electric Micromotor Dynamics,
IEEE Trans. Electron Devices, 39, 566-575
3. Mehregany, M., Senturia, S.D., Lang, J.H., Nagarkar, P. (1992) Micromotor Fabrication, IEEE Trans.
Electron Devices, 39, 2060-2069
4. Pacholik, J., Napieralski A., Grecki, M., Turowski, M. (1994) Temperature computation in semiconductor
devices using 3D network concept, Proc. of PEED, SPEEDAM'94, Taormina, Italy, Appendix 1-4
5. Turowski, M., Jablonski, G., Napieralski, A, Wiak, S. (1995) Electromechanical and Thermal Modelling of
IC-processed Micromotors, Proc. of Und Workshop: MIXED-VLSI DESIGN, Krakow, Poland, 343-348
6. Pacholik, J., Furmanczyk, M., Jablonski, G., Turowski M., Grecki M., Napieralski A. (1995)
Thermomechanical Design of Silicon Micropump", Proc. of Und Workshop: MIXED-VLSI DESIGN, Krakow,
Poland, 373 - 378
7. Furmaiiczyk M., Jablonski, G, Napieralski, A, Pacholik, J. (1995) The 3D transient thermal simulation with
arbitrary border conditions, Proc. of International THERMINIC Workshop, Grenoble 25-26/09/1995
8. Wöjciak, W., Napieralski A. (1995) The distance dependent method for temperature measurement of single
heat source in semiconductor devices - the first approach, Proc. of International THERMINIC Workshop,
Grenoble 25-26/09/1995
9. Napieralski A, Dorkel J.M., Leturcq Ph.,: "PYRTHERM - Manuel de l'utilisateur - Version 4.0 November
1990", rapport LAAS Nr. 90 380, (132 pages).
EMERGING AND FUTURE INTELLIGENT AVIATION & AUTOMOTIVE
APPLICATIONS OF MIMO ASIM MACROCOMMUTATORS AND ASIC
MICROCONTROLLERS
B.T. FIJALKOWSKI
Cracow University of Technology
Cracow, Poland
pmfijalk@cyf-kr. edu.pl
1. Introduction
Current trends in aviation and automotive macro-, meso-, micro- and nanoelectronics are
towards intelligent electronically-commutated (reciprocational and rotational) electro-
mechanical actuators (electrical machines) and other devices with much higher
performance and greater compactness, not only at the interconnection stage. The
increasing complexity of intelligent electronically-commutated electromechanical
actuators (electrical machines) and other devices is leading to larger chip sizes and the
greater degree of integration means that the power dissipated by the working chip is also
raising. Increasingly, the tendency is package several chips together in the form of
multichip module, using techniques such as wire bonding, tape automated bonding and
flip chip bonding to reduce the interconnections length and hence the overall size.
Continuing demands for high-performance intelligent multi-input/multi-output
(MIMO) macroelectronic AC-to-AC, AC-to-DC-to-AC, AC-to-DC/DC-to-AC,
DC-to-AC and DC-to-AC-to-DC converter commutators, called the AC-to-AC, AC-to-
DC-to-AC, AC-to-DC/DC-to-AC, DC-to-DC and DC-to-AC-to-DC macrocommutators,
and microelectronic neuro-fuzzy (NF) computer (processor) controllers, called the
NFmicrocontrollers are leading to integrated matrixers (IM) or application specific
integrated matrixers (ASIM) and integrated circuits (IC) or application specific integrated
circuits (ASIC) of far greater complexity than before. Artificial neural networks and
fuzzy logic (NF) are increasingly being incorporated in MIMO control systems to
provide robust and effective control in a wide range of aviation and automotive
applications. The method is based on a many-valued logic which enables general
principles and expert knowledge to be used to provide control rules and procedures. The
intrinsic non-linearity and variability of operating conditions make NF control on ideal
method for this area.
The author is actively pursuing the integration of matrixery and circuitry on the same
transparent substrate as the intelligent MIMO ASIM AC-to-AC, AC-to-DC-to-AC,
397
S. Luryi et al. (eds.), Future Trends in Microelectronics, 397-405.
© 1996 Khmer Academic Publishers. Printed in the Netherlands.
398
AC-to-DC/DC-to-AC, DC-to-DC and DC-to-AC-to-DC macrocommutators and ASIC
NF microcontrollers, respectively. The most obvious advantage of integrating matrixery
and circuitry on the same transparent substrate is the reduction in the 'work-horse'
components that are mono- and polycrystalline as well as amorphous-Si or GaAs super-
and/or semiconductor, fast-switching 'discrete' and/or 'continuous' uni- and/or bipolar
electrical valves [diodes, transistors, charge injection transistors (CHINTs), CMOS-
transistors, heterojunction bipolar transistors (HBTs), high electron mobility transistors
(HEMTs), insulated gate bipolar transistors (IGBTs), MES field effect transistors
(MESFETs), metal insulator-semiconductorFETs (MISFETs), MOSFETs, organicFETs
(OFETs), bipolar quantum interference transistors (QUITs), static induction transistors
(SITs), bipolar superconductor-base hot electron transistors (SUPER-HETs), gate-turn-
off (GTO) thyristors, light-triggered-and-quenched (LTQ) thyristors, MOS-controlled
thyristors (MCTs), ovonics etc], and in external connections.
The 'discrete' electrical valves suffer from having a performance that is strongly
dependent upon temperature. The author have therefore carried out a detailed program-
me of research into ASIMs using 'continuous' electrical valves, in which electrical
conduction is much less temperature-dependent than in 'discrete' electrical valves. The
external connections are also one of the dominating causes of ASIM and ASIC failure.
ASIMs and ASICs will be lighter and more compact because a proportion of the size and
mass of them. It is difficult to estimate the cost and reliability of these ASIMs and
ASICs as they are not yet in manufacture. Although research, development and initial
manufacture costs are high, the leading electronic manufacturers will be investigating
significantly in this key technology and the intelligent MIMO ASIM AC-to-AC,
AC-to-DC-to-AC, AC-to-DC/DC-to-AC , DC-to-DC and DC-to-AC-to-DC macro-
commutators capable of true inaudible (> 20 kHz) operation, and ASIC NF micro-
controllers will soon extend their application from the current aviation and automotive
intelligent electronically-commutated electromechanical actuators (electrical machines)
and other devices markets in to other areas and they have every chance of completely
revolutionizing the ASIM world.
The last few year have witnessed a rapid progress in macroelectronics (integrated
high-power electronics), mesoelectronics (integrated medium-power electronics), micro-
electronics (integrated low-power electronics) and nanoelectronics (integrated ultra-low-
power electronics), above all when it comes to the development that have taken place
in the field of avionics (aviation electronics) and automotronics (automotive electronics).
New applications for macro-, meso-, micro- and nanoelectronics have thus evolved and
established applications have undergone further advancements.
In 1973 I have forecasted a potential revolution in in applications of macro- and
microelectronicss as high-performance intelligent MIMO mono- and/or polycrystalline
as well as amorphous super- and/or semiconductor ASIM AC-to-AC, AC-to-DC-to-AC,
AC-to-DC/DC-to-AC, DC-to-DC and DC-to-AC-to-DC macrocommutators (conceived
and widely popularized by the author in the papers presented at the 1st Nat. RAILWAY
VEHICLES Conference, Krakow-Zawoja, Poland, October 1973; and the First European
Conference on POWER ELECTRONICS AND APPLICATIONS in Brussels, Belgium,
October 1984), and ASIC NF microcontrollers improve the economics of reshaping
399
electric power to accomplish various tasks not only in aviation and automotive sectors
but also in other applications.
Intelligent MIMO ASIM macrocommutators are the key to overcome the contactless
(brushless), ie sparkless commutation (changing over the way of electric-current flow),
and so to the static conversion of one kind of electrical energy into another, with the
application of the phenomena inducing uncontrolled or controlled electric current
conductivity (carrying of electric charges by positive or negative carriers into a definite
medium under the action of electric field).
Intelligent MIMO ASIM macrocommutators have announced the era of newly
designed energy-saving AC- and/or DC-powered elctromechanical actuators (electrical
machines), enabling continuous (stepless) and contactless control of their torque and
velocity or angular speed (during motoring), or their voltage and current (during
generating).
Intelligent MIMO ASIM macrocommutators with uni- and/or bipolar commutating
ASIMs constitute a successive breakthrough in the develiopment of AC- and/or DC-
powered electronically-commutated electromechanical actuators (electrical machines),
and multiply their possibilities enormously. There exist fantastic perspectives which are
probable to become reality as soon as at the end of this century. But not in this does the
scientific sense of concepts conceived by the author over twenty years ago consist.
The most essential value of this new concept was the achievement of cardinal
breakthrough in the thinking itself, on the static conversion of one kind of electrical
energy into another, with the aid of 'discrete' and/or 'continuous' uni- and/or bipolar
electrical valves - a change of paradigm, that was a kind of thinking standard in the field
of power electronics.
A.
t^tfiJ-H^ &
R-.-J&
£*k$<$-\-<J<£
C-
£$<f£$-i-<J<£
B-l.
^^£-H^£
&&&+<$<£
■1 n
7 7 7 7
zBn
EßA Z
BA ^BB ^BC ^B(N-l) ^BN ZB« ZBB Z* Zß(n -1) i«
7
ECB ZQA ZQB ZQC ZQN zCa Zcb zCc Zc(n -1) zCn iCB
p Z 7 ■N(N-
NA ZNB ZNC ZNN Zfia Zfib ZNC ZN(H -1) Zn„
(1)
Em 2*4zaB zaC 7
ZaN zaa z* zac • Zaip -1) z^ : an
■ba
ZbN zba z» Zu ■ zu
7
E* zM zbB z* • ^b(N-l) -1) z^
Ech z^ z* zcC 7
z* zca z* zcc • Z
c(n -1) zm •cb
(-«(n-l
p
^ znB Z„c
7
Z„N zm z^ z^ ■ z„(n -1) z
™.
Coefficients
BE.
MtH (,■'•('■-!) = 0; i*[) (2)
di'V-V
o-o-e—f-«3-e
"*J0,B-O-O \ O (P n-i,
AS
e-€)-ö
'CB\ CJI
e-e E
*.
IJ(S-l)
aB-&
aQ=gt=£)
N-li n-l
n(r.-l)
rs rs ! QUO
trs
E
ia *cb n(n-l)
Coefficients
k (4)
P\T
yW= d7_ (u. = 0; i*T)
du,
A
^0 CD 0 £>-e
Is.
Y^-Hy-^x-O ; O-O
-0-&-0 e-o W
—I-
nip
O-
|tJJ
IN-I -
1
1
**-e A*8]
i^ o o g>-e
■An
I ^L i*f
4. Conclusions
and
Bernard Courtois
TIMA, 46 Avenue Felix Viallet, 38031 Grenoble cedex, France
1. Introduction
In IC packaging DIL-like solutions were used for about twenty years. In the last
decade however new solutions are intensely investigated (QFP, PGA, flip-chip,
tape-bond, solder-bump, MCM, Cubic structures etc.). In all these the two main
goals are
- the highest possible pin number on the highest possible raster density,
— the best heat conductance.
We meet extreme dissipation values (e.g. 2000 W from a single MCM) and
unusual solutions for cooling (water cooling package etc.) [3]. Heat removal often
occurs directly from the active face to avoid even the thermal resistance contribu-
tion of the die itself. 3D packaging result in higher element density, consequently
even more serious thermal problems.
Thermal transient recording methods gain increasing role in the thermal inves-
tigation of IC structures and their packaging. In this method the thermal char-
acteristics of the structure is measured by a thermal step response recording. A
typical step-response function is shown in Fig. 2. This describes both the static
and the dynamic thermal behaviour. Many research papers deal with the problem
of finding methods for not only to indicate but also to locate heat sinking defects
based on these measurements. As a result of subsequent data processing the heat
flow map of the packaging structure can be obtained, on which the individual sec-
tions of the heat sinking path can be separated and the possible weak points or
anomalies can be located.
RECORD=PPoa5_$
Recorded: 11-18-1893 16:03:27
PPGrt package, soldered into PCboard
U
IE= 130.00 mA
UCBL= 1.000 U UCBH=10.000 U
ULSB= 0.02< mU alfa= 2.000 mU/K
Tfirst= 0.500 us Tmax=1800.000 s
nsample= 11718 sampl'oct= 500
i us:
Fig. 2. Thermal step-response function of an IC package. The z-axis is the time, and the y
axis is the temperature rise. Note, that the response has been obtained for an extremely wide
time range, from 1 ßs to 1000 s. The thermal time constants of the system are expected to be
distributed in this range
In the practice of the integrated circuit design it is usual to add some excess,
circuitry serving to support the design for testability (DfT). This method can be
411
Fig. 3. Measured temperature map of a small region on the IC surface. The result is obtained
ay liquid crystal method. The different colors (different shadings on this b-w picture) correspond
;o regions of different temperatures. The temperature step between two neighboring regions is
3.1 °C
3xtended for thermal testing. Recent experiments deal with the question of building
3ne or more thermal sensors into the chip and use them for thermal testing.
Chips supplied with thermal sensors can be thermally tested after fabrication
^DfTT, Design for Thermal Testability) [5]. With some additional circuitry the
temperature of the IC is readable also if the circuit is already built in an equip-
ment. This way the inner temperature of the IC can be controlled in the operating
equipment, preventing overheating and degradation. This is the basic principle of
thermal monitoring. The current problem of development is to find how can be
this problem solved with minimal extra expenses (chip area, excess pins). The
best solution seems to be the connection with the other electrical test circuits (e.g.
boundary scan) [6].
[n the case of delicate circuits the parameters can be improved if the circuit is
kept on constant temperature in a thermostat. The circuit itself can be designed
412
such that, it behaves as a kind of a micro thermostat. For this purpose a thermal
sensor and a dissipator are required on the chip surface, with some control circuit-
ry. The latter controls the dissipator to assure constant (e.g. 80 °C) chip surface
temperature. Such circuits are called TSS (temperature-stabilized surface) ICs.
In the case of higher dissipation circuits protection against overheating can
be useful. A temperature sensor observes the chip temperature and inhibits the
operation in the case of overheating, or reduces the dissipation e.g. in the case of
CMOS VLSI circuits by decreasing the clock frequency.
Exploiting the thermal effects different functional units can be realized in an inte-
grated circuit. As an example a dissipating resistor and a temperature gradient
sensor placed beside it form a true RMS (root-mean-square) measuring unit. A
dissipating element and the thermal sensor placed close to it can be used as a
flow meter. Based on similar principle the microelectronic version of the Pirani
vacuum meter can be constructed. Infrared sensor can be built by dropping IR
radiation on a (from the ambient thermally well isolated) target and measuring
the temperature increase of the target [7]. It has to be noted however that for the
fabrication of most of these units excess technological steps are required beyond
the usual IC process steps. At some places of the chip thin membranes or can-
tilevers have to be fabricated by etching, new surface layers of special material
are needed etc. These latter however belong already to another category: to the
integrated microsystems.
References
1. ANSYS - Engineering Analysis System: Theoretical Manual. Swanson Analysis Systems Inc.
2. V. Szekely and A. Poppe: Novel tools for thermal and electrical analysis of circuits. Electrosoft,
l(4):234-249, 1990. Computational Mechanics Publications.
3. R. E. Simons: The Evolution of IBM High Performance Cooling Technology. In Proc. of the
11th IEEE SEMI-THERM Symposium, pages 102-114, 7-9 Febr. 1995.
4. G. Gaussorgues: Infrared thermography. Chapman k. Hall, London, 1994.
5. V. Szekely et al: Design for Thermal Testability (DfTT) and a CMOS realisation. In Proc. of
the THERMINIC Workshop, Grenoble, France, 25-26 Sept 1995. (accepted for publication).
6. V. Szekely and M. Rencz: Thermal Test and Monitoring. In Proc. of the ED&T Conference
'95, Paris, page 601, 7-9 March 1995.
7. S. Middelhoek and S. A. Audet: Silicon Sensors. Academic Press, London, 1989.
CONTRIBUTORS
Michael A. Stroscio
US Army Research Office
P.O. Box 12211
Research Triangle,
NC 27709-2211
U.S.A.
Robert Suris
A.F. Ioffe Physico-Tech. Inst.
Polytechnicheskaya 26
St. Petersburg,
Russia
Index