Suchergebnis: Katalogdaten im Herbstsemester 2019
Doktorat Departement Informationstechnologie und Elektrotechnik ![]() Mehr Informationen unter: https://www.ethz.ch/de/doktorat.html | ||||||
![]() A minimum of 12 ECTS credit points must be obtained during doctoral studies. The courses on offer below are only a small selection out of a much larger available number of courses. Please discuss your course selection with your PhD supervisor. | ||||||
Nummer | Titel | Typ | ECTS | Umfang | Dozierende | |
---|---|---|---|---|---|---|
227-0225-00L | Linear System Theory | W | 6 KP | 5G | J. Lygeros | |
Kurzbeschreibung | The class is intended to provide a comprehensive overview of the theory of linear dynamical systems, stability analysis, and their use in control and estimation. The focus is on the mathematics behind the physical properties of these systems and on understanding and constructing proofs of properties of linear control systems. | |||||
Lernziel | Students should be able to apply the fundamental results in linear system theory to analyze and control linear dynamical systems. | |||||
Inhalt | - Proof techniques and practices. - Linear spaces, normed linear spaces and Hilbert spaces. - Ordinary differential equations, existence and uniqueness of solutions. - Continuous and discrete-time, time-varying linear systems. Time domain solutions. Time invariant systems treated as a special case. - Controllability and observability, duality. Time invariant systems treated as a special case. - Stability and stabilization, observers, state and output feedback, separation principle. | |||||
Skript | Available on the course Moodle platform. | |||||
Voraussetzungen / Besonderes | Sufficient mathematical maturity, in particular in linear algebra, analysis. | |||||
227-0377-00L | Physics of Failure and Failure Analysis of Electronic Devices and Equipment | W | 3 KP | 2V | U. Sennhauser | |
Kurzbeschreibung | Failures have to be avoided by proper design, material selection and manufacturing. Properties, degradation mechanisms, and expected lifetime of materials are introduced and the basics of failure analysis and analysis equipment are presented. Failures will be demonstrated experimentally and the opportunity is offered to perform a failure analysis with advanced equipment in the laboratory. | |||||
Lernziel | Introduction to the degradation and failure mechanisms and causes of electronic components, devices and systems as well as to methods and tools of reliability testing, characterization and failure analysis. | |||||
Inhalt | Summary of reliability and failure analysis terminology; physics of failure: materials properties, physical processes and failure mechanisms; failure analysis of ICs, PCBs, opto-electronics, discrete and other components and devices; basics and properties of instruments; application in circuit design and reliability analysis | |||||
Skript | Comprehensive copy of transparencies | |||||
227-0417-00L | Information Theory I | W | 6 KP | 4G | A. Lapidoth | |
Kurzbeschreibung | This course covers the basic concepts of information theory and of communication theory. Topics covered include the entropy rate of a source, mutual information, typical sequences, the asymptotic equi-partition property, Huffman coding, channel capacity, the channel coding theorem, the source-channel separation theorem, and feedback capacity. | |||||
Lernziel | The fundamentals of Information Theory including Shannon's source coding and channel coding theorems | |||||
Inhalt | The entropy rate of a source, Typical sequences, the asymptotic equi-partition property, the source coding theorem, Huffman coding, Arithmetic coding, channel capacity, the channel coding theorem, the source-channel separation theorem, feedback capacity | |||||
Literatur | T.M. Cover and J. Thomas, Elements of Information Theory (second edition) | |||||
227-0427-00L | Signal Analysis, Models, and Machine Learning | W | 6 KP | 4G | H.‑A. Loeliger | |
Kurzbeschreibung | Mathematical methods in signal processing and machine learning. I. Linear signal representation and approximation: Hilbert spaces, LMMSE estimation, regularization and sparsity. II. Learning linear and nonlinear functions and filters: neural networks, kernel methods. III. Structured statistical models: hidden Markov models, factor graphs, Kalman filter, Gaussian models with sparse events. | |||||
Lernziel | The course is an introduction to some basic topics in signal processing and machine learning. | |||||
Inhalt | Part I - Linear Signal Representation and Approximation: Hilbert spaces, least squares and LMMSE estimation, projection and estimation by linear filtering, learning linear functions and filters, L2 regularization, L1 regularization and sparsity, singular-value decomposition and pseudo-inverse, principal-components analysis. Part II - Learning Nonlinear Functions: fundamentals of learning, neural networks, kernel methods. Part III - Structured Statistical Models and Message Passing Algorithms: hidden Markov models, factor graphs, Gaussian message passing, Kalman filter and recursive least squares, Monte Carlo methods, parameter estimation, expectation maximization, linear Gaussian models with sparse events. | |||||
Skript | Lecture notes. | |||||
Voraussetzungen / Besonderes | Prerequisites: - local bachelors: course "Discrete-Time and Statistical Signal Processing" (5. Sem.) - others: solid basics in linear algebra and probability theory | |||||
227-0559-10L | Seminar in Communication Networks: Learning, Reasoning and Control ![]() Number of participants limited to 24. | W | 2 KP | 2S | L. Vanbever | |
Kurzbeschreibung | In this seminar participating students review, present, and discuss (mostly recent) research papers in the area of computer networks. During the fall semester of 2019, the seminar will focus on topics blending networks with machine learning and control theory. | |||||
Lernziel | The two main goals of this seminar are: 1) learning how to read and review scientific papers; and 2) learning how to present and discuss technical topics with an audience of peers. Students are required to attend the entire seminar, choose a paper to present from a given list, prepare and give a presentation on that topic, and lead the follow-up discussion. To ensure the talks' quality, each student will be mentored by a teaching assistant. In addition to presenting one paper, every student is also required to submit one (short) review for one of the two papers presented every week in-class (12 reviews in total). The students will be evaluated based on their submitted reviews, their presentation, their leadership in animating the discussion for their own paper, and their participation in the discussions of other papers. | |||||
Inhalt | The seminar will start with two introductory lectures in week 1 and week 2. Starting from week 3, participating students will start reviewing, presenting, and discussing research papers. Each week will see two presentations, for a total of 24 papers. The course content will vary from semester to semester. During the fall semester of 2019, the seminar will focus on topics blending networks with machine learning and control theory. For details, please see: https://seminar-net.ethz.ch | |||||
Skript | The slides of each presentation will be made available on the website. | |||||
Literatur | The paper selection will be made available on the course website: https://seminar-net.ethz.ch | |||||
Voraussetzungen / Besonderes | Communication Networks (227-0120-00L) or equivalents. For fall 2019, it is expected that students have prior knowledge in machine learning and control theory, for instance by having attended appropriate courses. | |||||
227-0689-00L | System Identification | W | 4 KP | 2V + 1U | R. Smith | |
Kurzbeschreibung | Theory and techniques for the identification of dynamic models from experimentally obtained system input-output data. | |||||
Lernziel | To provide a series of practical techniques for the development of dynamical models from experimental data, with the emphasis being on the development of models suitable for feedback control design purposes. To provide sufficient theory to enable the practitioner to understand the trade-offs between model accuracy, data quality and data quantity. | |||||
Inhalt | Introduction to modeling: Black-box and grey-box models; Parametric and non-parametric models; ARX, ARMAX (etc.) models. Predictive, open-loop, black-box identification methods. Time and frequency domain methods. Subspace identification methods. Optimal experimental design, Cramer-Rao bounds, input signal design. Parametric identification methods. On-line and batch approaches. Closed-loop identification strategies. Trade-off between controller performance and information available for identification. | |||||
Literatur | "System Identification; Theory for the User" Lennart Ljung, Prentice Hall (2nd Ed), 1999. "Dynamic system identification: Experimental design and data analysis", GC Goodwin and RL Payne, Academic Press, 1977. | |||||
Voraussetzungen / Besonderes | Control systems (227-0216-00L) or equivalent. | |||||
227-0955-00L | Seminar in Electromagnetics, Photonics and Terahertz ![]() | W | 3 KP | 2S | J. Leuthold | |
Kurzbeschreibung | Selected topics of the current research activities at the IEF and closely related institutions are discussed. | |||||
Lernziel | Have an overview on the research activities of the IEF institute. | |||||
227-0974-00L | TNU Colloquium ![]() | W | 0 KP | 2K | K. Stephan | |
Kurzbeschreibung | This colloquium for MSc/PhD students at D-ITET discusses current research in Neuromodeling (the development of mathematical models for diagnostics of brain diseases) and application to Computational Psychiatry/Psychosomatics. The range of topics is broad, incl. statistics and computational modeling, experimental paradigms (fMRI, EEG, behaviour), and clinical questions. | |||||
Lernziel | see above | |||||
252-0535-00L | Advanced Machine Learning ![]() | W | 8 KP | 3V + 2U + 2A | J. M. Buhmann | |
Kurzbeschreibung | Machine learning algorithms provide analytical methods to search data sets for characteristic patterns. Typical tasks include the classification of data, function fitting and clustering, with applications in image and speech analysis, bioinformatics and exploratory data analysis. This course is accompanied by practical machine learning projects. | |||||
Lernziel | Students will be familiarized with advanced concepts and algorithms for supervised and unsupervised learning; reinforce the statistics knowledge which is indispensible to solve modeling problems under uncertainty. Key concepts are the generalization ability of algorithms and systematic approaches to modeling and regularization. Machine learning projects will provide an opportunity to test the machine learning algorithms on real world data. | |||||
Inhalt | The theory of fundamental machine learning concepts is presented in the lecture, and illustrated with relevant applications. Students can deepen their understanding by solving both pen-and-paper and programming exercises, where they implement and apply famous algorithms to real-world data. Topics covered in the lecture include: Fundamentals: What is data? Bayesian Learning Computational learning theory Supervised learning: Ensembles: Bagging and Boosting Max Margin methods Neural networks Unsupservised learning: Dimensionality reduction techniques Clustering Mixture Models Non-parametric density estimation Learning Dynamical Systems | |||||
Skript | No lecture notes, but slides will be made available on the course webpage. | |||||
Literatur | C. Bishop. Pattern Recognition and Machine Learning. Springer 2007. R. Duda, P. Hart, and D. Stork. Pattern Classification. John Wiley & Sons, second edition, 2001. T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical Learning: Data Mining, Inference and Prediction. Springer, 2001. L. Wasserman. All of Statistics: A Concise Course in Statistical Inference. Springer, 2004. | |||||
Voraussetzungen / Besonderes | The course requires solid basic knowledge in analysis, statistics and numerical methods for CSE as well as practical programming experience for solving assignments. Students should have followed at least "Introduction to Machine Learning" or an equivalent course offered by another institution. PhD students are required to obtain a passing grade in the course (4.0 or higher based on project and exam) to gain credit points. | |||||
252-0417-00L | Randomized Algorithms and Probabilistic Methods | W | 8 KP | 3V + 2U + 2A | A. Steger | |
Kurzbeschreibung | Las Vegas & Monte Carlo algorithms; inequalities of Markov, Chebyshev, Chernoff; negative correlation; Markov chains: convergence, rapidly mixing; generating functions; Examples include: min cut, median, balls and bins, routing in hypercubes, 3SAT, card shuffling, random walks | |||||
Lernziel | After this course students will know fundamental techniques from probabilistic combinatorics for designing randomized algorithms and will be able to apply them to solve typical problems in these areas. | |||||
Inhalt | Randomized Algorithms are algorithms that "flip coins" to take certain decisions. This concept extends the classical model of deterministic algorithms and has become very popular and useful within the last twenty years. In many cases, randomized algorithms are faster, simpler or just more elegant than deterministic ones. In the course, we will discuss basic principles and techniques and derive from them a number of randomized methods for problems in different areas. | |||||
Skript | Yes. | |||||
Literatur | - Randomized Algorithms, Rajeev Motwani and Prabhakar Raghavan, Cambridge University Press (1995) - Probability and Computing, Michael Mitzenmacher and Eli Upfal, Cambridge University Press (2005) | |||||
263-4500-00L | Advanced Algorithms ![]() | W | 6 KP | 2V + 2U + 1A | M. Ghaffari, A. Krause | |
Kurzbeschreibung | This is an advanced course on the design and analysis of algorithms, covering a range of topics and techniques not studied in typical introductory courses on algorithms. | |||||
Lernziel | This course is intended to familiarize students with (some of) the main tools and techniques developed over the last 15-20 years in algorithm design, which are by now among the key ingredients used in developing efficient algorithms. | |||||
Inhalt | The lectures will cover a range of topics, including the following: graph sparsifications while preserving cuts or distances, various approximation algorithms techniques and concepts, metric embeddings and probabilistic tree embeddings, online algorithms, multiplicative weight updates, streaming algorithms, sketching algorithms. | |||||
Skript | https://people.inf.ethz.ch/gmohsen/AA19/ | |||||
Voraussetzungen / Besonderes | This course is designed for masters and doctoral students and it especially targets those interested in theoretical computer science, but it should also be accessible to last-year bachelor students. Sufficient comfort with both (A) Algorithm Design & Analysis and (B) Probability & Concentrations. E.g., having passed the course Algorithms, Probability, and Computing (APC) is highly recommended, though not required formally. If you are not sure whether you're ready for this class or not, please consult the instructor. | |||||
327-2132-00L | Multifunctional Ferroic Materials: Growth, Characterisation, Simulation | W | 2 KP | 2G | M. Trassin, M. Fiebig | |
Kurzbeschreibung | The course will explore the growth of (multi-) ferroic oxide thin films. The structural characterization and ferroic state investigation by force microscopy and by laser-optical techniques will be addressed. Oxide electronics device concepts will be discussed. | |||||
Lernziel | Oxide films with a thickness of just a few atoms can now be grown with a precision matching that of semiconductors. This opens up a whole world of functional device concepts and fascinating phenomena that would not occur in the expanded bulk crystal. Particularly interesting phenomena occur in films showing magnetic or electric order or, even better, both of these ("multiferroics"). In this course students will obtain an overarching view on oxide thin epitaxial films and heterostructures design, reaching from their growth by pulsed laser deposition to an understanding of their magnetoelectric functionality from advanced characterization techniques. Students will therefore understand how to fabricate and characterize highly oriented films with magnetic and electric properties not found in nature. | |||||
Inhalt | Types of ferroic order, multiferroics, oxide materials, thin-film growth by pulsed laser deposition, molecular beam epitaxy, RF sputtering, structural characterization (reciprocal space - basics-, XRD for thin films, RHEED) epitaxial strain related effects, scanning probe microscopy techniques, laser-optical characterization, oxide thin film based devices and examples. | |||||
401-3055-64L | Algebraic Methods in Combinatorics ![]() | W | 6 KP | 2V + 1U | B. Sudakov | |
Kurzbeschreibung | Combinatorics is a fundamental mathematical discipline as well as an essential component of many mathematical areas, and its study has experienced an impressive growth in recent years. This course provides a gentle introduction to Algebraic methods, illustrated by examples and focusing on basic ideas and connections to other areas. | |||||
Lernziel | The students will get an overview of various algebraic methods for solving combinatorial problems. We expect them to understand the proof techniques and to use them autonomously on related problems. | |||||
Inhalt | Combinatorics is a fundamental mathematical discipline as well as an essential component of many mathematical areas, and its study has experienced an impressive growth in recent years. While in the past many of the basic combinatorial results were obtained mainly by ingenuity and detailed reasoning, the modern theory has grown out of this early stage and often relies on deep, well-developed tools. One of the main general techniques that played a crucial role in the development of Combinatorics was the application of algebraic methods. The most fruitful such tool is the dimension argument. Roughly speaking, the method can be described as follows. In order to bound the cardinality of of a discrete structure A one maps its elements to vectors in a linear space, and shows that the set A is mapped to linearly independent vectors. It then follows that the cardinality of A is bounded by the dimension of the corresponding linear space. This simple idea is surprisingly powerful and has many famous applications. This course provides a gentle introduction to Algebraic methods, illustrated by examples and focusing on basic ideas and connections to other areas. The topics covered in the class will include (but are not limited to): Basic dimension arguments, Spaces of polynomials and tensor product methods, Eigenvalues of graphs and their application, the Combinatorial Nullstellensatz and the Chevalley-Warning theorem. Applications such as: Solution of Kakeya problem in finite fields, counterexample to Borsuk's conjecture, chromatic number of the unit distance graph of Euclidean space, explicit constructions of Ramsey graphs and many others. The course website can be found at https://moodle-app2.let.ethz.ch/course/view.php?id=11617 | |||||
Skript | Lectures will be on the blackboard only, but there will be a set of typeset lecture notes which follow the class closely. | |||||
Voraussetzungen / Besonderes | Students are expected to have a mathematical background and should be able to write rigorous proofs. | |||||
401-5680-00L | Foundations of Data Science Seminar ![]() | Z | 0 KP | P. L. Bühlmann, A. Bandeira, H. Bölcskei, J. M. Buhmann, T. Hofmann, A. Krause, A. Lapidoth, H.‑A. Loeliger, M. H. Maathuis, G. Rätsch, C. Uhler, S. van de Geer | ||
Kurzbeschreibung | Research colloquium | |||||
Lernziel | ||||||
» Auswahl aus sämtlichen Lehrveranstaltungen der ETH Zürich |
Seite 1 von 1