CoDaDri2: Difference between revisions

From Wiki Cours
Jump to navigation Jump to search
Line 83: Line 83:


* Introductory notebooks: [https://colab.research.google.com/drive/1Ovrh1JjLzMMnxtasDgR1j84CHGU6hMxA python], [https://colab.research.google.com/drive/14PWu-C171pYhV8ven-Txc4wz5Tv1DCsK numpy] and [https://colab.research.google.com/drive/1fPjwrdxQPSyfXniyAHmyOcF_gK2fqjts matplotlib]
* Introductory notebooks: [https://colab.research.google.com/drive/1Ovrh1JjLzMMnxtasDgR1j84CHGU6hMxA python], [https://colab.research.google.com/drive/14PWu-C171pYhV8ven-Txc4wz5Tv1DCsK numpy] and [https://colab.research.google.com/drive/1fPjwrdxQPSyfXniyAHmyOcF_gK2fqjts matplotlib]


* [https://colab.research.google.com/drive/1g7HXFUBQUBF0fhy5h2jYbfIAVBTqvb7-#scrollTo=U9LFX8OFhHOG Homework 1] (deadline October 1)
* [https://colab.research.google.com/drive/1g7HXFUBQUBF0fhy5h2jYbfIAVBTqvb7-#scrollTo=U9LFX8OFhHOG Homework 1] (deadline October 1)
Line 91: Line 90:
* [https://colab.research.google.com/drive/1k60-ChM3aUWjGsiRsf2Idx9jxHHtwyAb?usp=sharing Tutorial 3] Thumb rule [https://colab.research.google.com/drive/1fdKJlp0lD4k530oRuu50_Hm_T8t3jixa?usp=sharing problems]
* [https://colab.research.google.com/drive/1k60-ChM3aUWjGsiRsf2Idx9jxHHtwyAb?usp=sharing Tutorial 3] Thumb rule [https://colab.research.google.com/drive/1fdKJlp0lD4k530oRuu50_Hm_T8t3jixa?usp=sharing problems]
-->
-->


'''Friday, September 16, 2022'''
'''Friday, September 16, 2022'''


* [https://drive.google.com/file/d/1eY5cuNcJqYw7df0PBkxq03tEnqzTOC-5/view?usp=sharing Lecture 3 ]: Importance sampling
* [https://drive.google.com/file/d/1eY5cuNcJqYw7df0PBkxq03tEnqzTOC-5/view?usp=sharing Lecture 3 ]: Importance sampling
Line 129: Line 125:
'''Friday, October 7, 2022'''
'''Friday, October 7, 2022'''


* Test: XXX
* Test:  
 
The MCQ is composed of 19 questions (one of them counts for two).  For each question you have 4 choices: 3 wrong and 1 correct: If you check the correct one you get a point. If you are wrong you loose 1/4 of a point.  No answer given: zero points.




Line 137: Line 135:
* [https://colab.research.google.com/drive/1KyT442vB4EuGKBy_Xj3j8cLtMMl4gbac?usp=sharing Tutorial 6]: Faster than the clock algorithms [https://colab.research.google.com/drive/1C-ieAwaAKwkX5MFm8aX2kNO2GNeO9CdY?usp=sharing problems]
* [https://colab.research.google.com/drive/1KyT442vB4EuGKBy_Xj3j8cLtMMl4gbac?usp=sharing Tutorial 6]: Faster than the clock algorithms [https://colab.research.google.com/drive/1C-ieAwaAKwkX5MFm8aX2kNO2GNeO9CdY?usp=sharing problems]
-->
-->


'''Friday, October 14, 2022'''
'''Friday, October 14, 2022'''


* Lecture 7: Introduction to Bayesian inference
* Lecture 7: Introduction to Bayesian inference
Line 149: Line 142:
* Tutorial 7: Bayesian inference and single-particle tracking. [https://drive.google.com/file/d/1q9Quq5mvJ4Avv2qWgzTocU-DVIxJIVWi/view?usp=sharing Questions]. [https://drive.google.com/file/d/1xYb4wfNBysse6Q7frTw_V9TPdUQU5lUx/view?usp=sharing Data]. [https://drive.google.com/file/d/1Ba_kRo3_XMIaDLNYZlbZEw-wGXKS-29I/view?usp=sharing Starting Notebook]. [https://colab.research.google.com/drive/1VePL9X1JT6yl8LmROeZyMyf4Epcu3yOL#scrollTo=1XQNGa_fyEkJ Google colab version]. Solutions [https://drive.google.com/file/d/1Q9IjVUeL7AA8uWk8DX9Vy9vuUQ21RYh2/view?usp=sharing]. Notebook [https://drive.google.com/file/d/1ZJmpdaTQVUttEaaI1ch-reFfZGI93454/view?usp=sharing]
* Tutorial 7: Bayesian inference and single-particle tracking. [https://drive.google.com/file/d/1q9Quq5mvJ4Avv2qWgzTocU-DVIxJIVWi/view?usp=sharing Questions]. [https://drive.google.com/file/d/1xYb4wfNBysse6Q7frTw_V9TPdUQU5lUx/view?usp=sharing Data]. [https://drive.google.com/file/d/1Ba_kRo3_XMIaDLNYZlbZEw-wGXKS-29I/view?usp=sharing Starting Notebook]. [https://colab.research.google.com/drive/1VePL9X1JT6yl8LmROeZyMyf4Epcu3yOL#scrollTo=1XQNGa_fyEkJ Google colab version]. Solutions [https://drive.google.com/file/d/1Q9IjVUeL7AA8uWk8DX9Vy9vuUQ21RYh2/view?usp=sharing]. Notebook [https://drive.google.com/file/d/1ZJmpdaTQVUttEaaI1ch-reFfZGI93454/view?usp=sharing]


 
'''Friday, October 21, 2022'''
 
 
 
'''Friday, October 22, 2021'''


* Lecture 8: Asymptotic inference and information. Extra material: Proof of Cramer-Rao bound [https://drive.google.com/file/d/10Ph_iP6AIQ3ps9v3FwRBO9j6-qV7oKV9/view?usp=sharing]
* Lecture 8: Asymptotic inference and information. Extra material: Proof of Cramer-Rao bound [https://drive.google.com/file/d/10Ph_iP6AIQ3ps9v3FwRBO9j6-qV7oKV9/view?usp=sharing]
Line 161: Line 150:
* Send your copy of Homework 2 to numphys.icfp  at  gmail.com  Thanks!
* Send your copy of Homework 2 to numphys.icfp  at  gmail.com  Thanks!


 
'''Friday, October 28, 2022'''
'''Friday, October 29, 2021'''


* Lecture 9: High-dimensional inference and Principal Component Analysis. Extra material: Handwritten notes on the derivation of Marcenko-Pastur spectral density [https://drive.google.com/file/d/10Yytv9itdWDDHTsM2lQHwQDMbcf-oVnd/view?usp=sharing]
* Lecture 9: High-dimensional inference and Principal Component Analysis. Extra material: Handwritten notes on the derivation of Marcenko-Pastur spectral density [https://drive.google.com/file/d/10Yytv9itdWDDHTsM2lQHwQDMbcf-oVnd/view?usp=sharing]
Line 169: Line 157:
[https://drive.google.com/file/d/11E75QwrJgDKrutGj9t5NJTRkGtuRgoBI/view?usp=sharing Solutions]. [https://drive.google.com/file/d/1-39oZNNB-35zM4V2tLWtubQnZTdMLtIz/view?usp=sharing Notebook]
[https://drive.google.com/file/d/11E75QwrJgDKrutGj9t5NJTRkGtuRgoBI/view?usp=sharing Solutions]. [https://drive.google.com/file/d/1-39oZNNB-35zM4V2tLWtubQnZTdMLtIz/view?usp=sharing Notebook]


 
'''Friday, November 25, 2022'''
 
'''Friday, November 12, 2021, 2 pm: The Quiz.'''
 
The MCQ is composed of 19 questions (one of them counts for two).  For each question you have 4 choices: 3 wrong and 1 correct: If you check the correct one you get a point. If you are wrong you loose 1/4 of a point.  No answer given: zero points.
 
 
MCQ Solution (correct answers in bold):[https://drive.google.com/file/d/1WWcqxFzmCJp2ZXC02a70a1o17pGupdyZ/view?usp=sharing]
 
 
'''Friday, November 26, 2021'''


* Lecture 10: Priors, regularisation, sparsity
* Lecture 10: Priors, regularisation, sparsity
Line 186: Line 164:
[https://drive.google.com/file/d/161ZLuq5s2RpHDJWELFSKXqlf-qwrVrxA/view?usp=sharing Notebook on Artificial data] [https://drive.google.com/file/d/17qhYN8EFpcoCUHOM_juBvEGpu61GgJRx/view?usp=sharing Corrections ]
[https://drive.google.com/file/d/161ZLuq5s2RpHDJWELFSKXqlf-qwrVrxA/view?usp=sharing Notebook on Artificial data] [https://drive.google.com/file/d/17qhYN8EFpcoCUHOM_juBvEGpu61GgJRx/view?usp=sharing Corrections ]


 
'''Friday, December 2, 2022'''
 
 
'''Friday, December 3, 2021'''


* Lecture 11: Probabilistic graphical models
* Lecture 11: Probabilistic graphical models
Line 195: Line 170:
* Tutorial 11:[https://drive.google.com/file/d/1A5pcJICHkmHZYoa2uoAGSAWp-_ZOaIEf/view?usp=sharing Analysis of protein sequence data to infer protein structure] [https://drive.google.com/file/d/1COKr5pNoBRFwwnj7TWPQSU8mEWKvsLud/view?usp=sharing Starting notebook and data] [https://drive.google.com/file/d/1CQw1PQ6RSS6nuGxhkwOqe7LJHpOPIaVp/view?usp=sharing Biblio][https://drive.google.com/file/d/1CslX27bTp5gyhXV4ciFE8s1zMnBgJ-1O/view?usp=sharing Solutions] [https://drive.google.com/file/d/19X59x3TdIsJaccTZE71gCbZxFbOzlhj4/view?usp=sharing Final notebook]
* Tutorial 11:[https://drive.google.com/file/d/1A5pcJICHkmHZYoa2uoAGSAWp-_ZOaIEf/view?usp=sharing Analysis of protein sequence data to infer protein structure] [https://drive.google.com/file/d/1COKr5pNoBRFwwnj7TWPQSU8mEWKvsLud/view?usp=sharing Starting notebook and data] [https://drive.google.com/file/d/1CQw1PQ6RSS6nuGxhkwOqe7LJHpOPIaVp/view?usp=sharing Biblio][https://drive.google.com/file/d/1CslX27bTp5gyhXV4ciFE8s1zMnBgJ-1O/view?usp=sharing Solutions] [https://drive.google.com/file/d/19X59x3TdIsJaccTZE71gCbZxFbOzlhj4/view?usp=sharing Final notebook]


 
'''Friday, December 9, 2021'''
'''Friday, December 10, 2021'''


* Lecture 12: Hidden Markov Models. Extra material: Pedagogical introduction to Kalman filters [https://drive.google.com/file/d/0B4HvoS7Zt11LZUxVLTkyVXRQRUE/view?usp=sharing&resourcekey=0-CzsYaRIrPP2sN-UmwfiLHQ]
* Lecture 12: Hidden Markov Models. Extra material: Pedagogical introduction to Kalman filters [https://drive.google.com/file/d/0B4HvoS7Zt11LZUxVLTkyVXRQRUE/view?usp=sharing&resourcekey=0-CzsYaRIrPP2sN-UmwfiLHQ]
Line 203: Line 177:
[https://drive.google.com/file/d/1FLbWZRWJ1JILZV41RjwVKR7mOhNMhLV9/view?usp=sharing Hidden Markov Models Hidden  for identification of recombinations in SARS-CoV-2 viral genomes] [https://drive.google.com/file/d/1FCAg0ihMWoAk-_dtt5ORncjVjkVu-m4D/view?usp=sharing Starting Notebook and Data][https://drive.google.com/file/d/1HMREe6ge7_K-lOKC4f2FwL7vnEa1_8Yf/view?usp=sharing Bibliography] [https://drive.google.com/file/d/1FLDBJmNtXogHiO5VfScGiePTCxMFy1KZ/view?usp=sharing Final Notebook] [https://drive.google.com/file/d/1FBQaNJ2RhjwW_VSut2oC7mZYl5F8RWfS/view?usp=sharing Solutions]
[https://drive.google.com/file/d/1FLbWZRWJ1JILZV41RjwVKR7mOhNMhLV9/view?usp=sharing Hidden Markov Models Hidden  for identification of recombinations in SARS-CoV-2 viral genomes] [https://drive.google.com/file/d/1FCAg0ihMWoAk-_dtt5ORncjVjkVu-m4D/view?usp=sharing Starting Notebook and Data][https://drive.google.com/file/d/1HMREe6ge7_K-lOKC4f2FwL7vnEa1_8Yf/view?usp=sharing Bibliography] [https://drive.google.com/file/d/1FLDBJmNtXogHiO5VfScGiePTCxMFy1KZ/view?usp=sharing Final Notebook] [https://drive.google.com/file/d/1FBQaNJ2RhjwW_VSut2oC7mZYl5F8RWfS/view?usp=sharing Solutions]


 
'''Friday, December 16, 2021'''
 
'''Friday, December 17, 2021'''


* Lecture 13: Unsupervised learning and representations
* Lecture 13: Unsupervised learning and representations

Revision as of 15:21, 22 June 2022

Breaking news:

  • Update with course schedule

Computational and Data Driven Physics

Modern physics is characterized by an increasing complexity of systems under investigation, in domains as diverse as condensed matter, astrophysics, biophysics, etc. Establishing adequate models to describe these systems and being able to make quantitative predictions from those models is extremely challenging. The goal of the course is to provide the tools and concepts necessary to tackle those systems.

Course description

We will first cover many algorithms used in many-body problems and complex systems, with special emphasis on Monte Carlo methods, molecular dynamics, and optimization in complex landscapes.

Second, we will provide statistical inference and machine learning tools to harness the growing availability of experimental data to design accurate models of the underlying, complex, strongly non-homogeneous and interacting systems.

Each theoretical lecture will be followed by a tutorial illustrating the concepts with practical applications borrowed from various domains of physics. We will focus on methods and algorithms and physics, not on programming and heavy numerics! You will have to hand in 3 homeworks.

The Team

Where and When

  • Lectures on Fridays: 14:00-16:00
  • Tutorials on Fridays: 16:00-18:00
  • ENS, 29 rue D'Ulm, salle Borel + Djebar

Slack

If you have questions or want to discuss topics related to the lecture, to the exercises or to the homeworks, you can use the [XXX Computational and Data Driven Physics Slack]. In order to join the Slack use the following [XXX invitation link].

Computer Requirements

No previous experience in programming is required.

Programming Language: Python

For practical installation, we recommand either to use Anaconda (See Memento Python) or use google colab.
The Collaboratory platform from Google is quite good way to use powerful computer without buying one: It requires no specific hardware or software, and even allows you to use GPU computing for free, all by writting a jupyter notebook that you can then share.

Grading

Computational Physics:

  • Homework 1: 5 points
  • Homework 2: 5 points
  • Multiple Choice Questions in November: 10 points



Data Driven Physics:

  • Homework 3: 5 points
  • Final exam in January: 15 points

Schedule

Friday, September 2, 2022


Friday, September 9, 2022


Friday, September 16, 2022


Friday, September 23, 2022


Friday, September 30, 2022

  • Send your copy of Homework 1 to numphys.icfp at gmail.com Thanks!


Friday, October 7, 2022

  • Test:

The MCQ is composed of 19 questions (one of them counts for two). For each question you have 4 choices: 3 wrong and 1 correct: If you check the correct one you get a point. If you are wrong you loose 1/4 of a point. No answer given: zero points.



Friday, October 14, 2022

  • Lecture 7: Introduction to Bayesian inference

Friday, October 21, 2022

  • Lecture 8: Asymptotic inference and information. Extra material: Proof of Cramer-Rao bound [3]
  • Send your copy of Homework 2 to numphys.icfp at gmail.com Thanks!

Friday, October 28, 2022

  • Lecture 9: High-dimensional inference and Principal Component Analysis. Extra material: Handwritten notes on the derivation of Marcenko-Pastur spectral density [4]

Solutions. Notebook

Friday, November 25, 2022

  • Lecture 10: Priors, regularisation, sparsity

Notebook on Artificial data Corrections

Friday, December 2, 2022

  • Lecture 11: Probabilistic graphical models

Friday, December 9, 2021

  • Lecture 12: Hidden Markov Models. Extra material: Pedagogical introduction to Kalman filters [5]
  • Tutorial 12:

Hidden Markov Models Hidden for identification of recombinations in SARS-CoV-2 viral genomes Starting Notebook and DataBibliography Final Notebook Solutions

Friday, December 16, 2021

  • Lecture 13: Unsupervised learning and representations


Final examination of the data-driven course (January 7, 2022)

  • Example of exam: On-line Principal Component Analysis [6]
  • On-line version of the book [7]
  • Examination repository [8]