Skip to content

Professional-grade presentations on advanced statistics, MCMC methods, and data science applications. Created by ESMAD faculty and Mysense.ai Lead Data Scientist. Includes theoretical foundations, modern algorithms, and industry implementations.

Notifications You must be signed in to change notification settings

DiogoRibeiro7/academic-presentations

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

31 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Academic Presentations

Comprehensive Data Science & Machine Learning Course Materials

Diogo Ribeiro
ESMAD - Escola Superior de MΓ©dia Arte e Design
Lead Data Scientist, Mysense.ai

ORCID Email Institution Company License: CC BY-SA 4.0 Code License: MIT

🎯 Overview

This repository contains a comprehensive collection of professional academic presentations covering advanced topics in statistics, machine learning, deep learning, and data science. The materials are designed for:

  • πŸŽ“ Graduate-level courses in Data Science, Statistics, and Computer Science
  • πŸ”¬ Research seminars and academic conferences
  • 🏒 Professional training programs in industry
  • πŸ“š Self-study for advanced learners

Key Features

βœ… 15+ comprehensive presentations with 100+ hours of content
βœ… Production-ready code in Python and R (27,000+ lines)
βœ… 140+ curated references with DOIs
βœ… Professional LaTeX theme with consistent styling
βœ… Hands-on exercises and assessments
βœ… Automated PDF generation via GitHub Actions

πŸ“š Course Catalog & Learning Objectives

πŸ”· Deep Learning & Neural Networks

Deep Learning Fundamentals

πŸ“‚ 02-deep-learning/deep-learning-fundamentals/

Learning Objectives:

  • Understand the mathematical foundations of neural networks
  • Implement backpropagation and gradient descent from scratch
  • Master modern optimization techniques (SGD, Adam, AdamW)
  • Design and train CNN architectures for computer vision
  • Build RNN/LSTM models for sequential data
  • Understand Transformer architecture and attention mechanisms
  • Apply regularization techniques (dropout, batch normalization)

Topics Covered:

  • Perceptron and multilayer networks
  • Activation functions (ReLU, sigmoid, tanh, Swish)
  • Loss functions and optimization
  • Convolutional Neural Networks (LeNet, AlexNet, VGG, ResNet)
  • Recurrent Neural Networks and LSTM
  • Transformers and self-attention
  • Training best practices

Prerequisites: Linear algebra, calculus, Python programming
Level: Intermediate to Advanced
Duration: 3-4 weeks (graduate course)


Reinforcement Learning

πŸ“‚ 02-deep-learning/reinforcement-learning/

Learning Objectives:

  • Formulate problems as Markov Decision Processes
  • Derive and apply Bellman equations
  • Implement value iteration and policy iteration
  • Understand Monte Carlo and TD learning methods
  • Build Q-learning and SARSA agents
  • Apply function approximation with neural networks
  • Implement modern deep RL algorithms (DQN, PPO, A3C)
  • Design multi-agent systems

Topics Covered:

  • Markov Decision Processes and dynamic programming
  • Monte Carlo methods
  • Temporal Difference learning (SARSA, Q-learning)
  • Function approximation and deep Q-networks
  • Policy gradient methods (REINFORCE, Actor-Critic, PPO)
  • Multi-agent reinforcement learning
  • Applications (games, robotics, resource allocation)

Prerequisites: Probability, linear algebra, Python
Level: Advanced
Duration: 4-5 weeks (graduate course)


πŸ”· Machine Learning Theory & Practice

Statistical Learning Theory

πŸ“‚ 01-foundations/statistical-modeling/

Learning Objectives:

  • Understand bias-variance tradeoff
  • Master regularization techniques (Ridge, Lasso, Elastic Net)
  • Apply cross-validation and model selection
  • Implement ensemble methods (bagging, boosting, stacking)
  • Understand kernel methods and SVMs
  • Perform dimensionality reduction (PCA, t-SNE, UMAP)
  • Evaluate models using appropriate metrics

Topics Covered:

  • Supervised learning fundamentals
  • Linear and logistic regression
  • Regularization and model selection
  • Tree-based methods (CART, Random Forests, XGBoost)
  • Support Vector Machines
  • Gaussian Processes
  • Model evaluation and validation

Prerequisites: Statistics, linear algebra, programming
Level: Intermediate
Duration: 4-5 weeks


Feature Engineering

πŸ“‚ 01-foundations/feature-engineering/

Learning Objectives:

  • Design effective feature engineering pipelines
  • Handle missing data with advanced imputation techniques
  • Encode categorical variables appropriately
  • Create polynomial and interaction features
  • Apply feature scaling and normalization
  • Perform feature selection using multiple methods
  • Build end-to-end ML pipelines with scikit-learn

Topics Covered:

  • Missing value imputation (mean, median, KNN, MICE)
  • Categorical encoding (one-hot, ordinal, target, entity embeddings)
  • Feature scaling (standard, min-max, robust)
  • Polynomial features and interactions
  • Feature selection (filter, wrapper, embedded methods)
  • Dimensionality reduction
  • Pipeline construction

Prerequisites: Basic Python, pandas, scikit-learn
Level: Beginner to Intermediate
Duration: 2-3 weeks


Explainable AI & Model Interpretability

πŸ“‚ 06-advanced-topics/explainable-ai/

Learning Objectives:

  • Understand the interpretability-accuracy tradeoff
  • Explain model predictions using SHAP values
  • Apply LIME for local explanations
  • Compute and interpret permutation importance
  • Visualize partial dependence and ICE plots
  • Detect and mitigate algorithmic bias
  • Implement fairness metrics and constraints
  • Use modern XAI tools (SHAP, LIME, InterpretML)

Topics Covered:

  • Global vs local explanations
  • Model-agnostic methods (SHAP, LIME, permutation importance)
  • Model-specific interpretability (linear models, trees, neural networks)
  • Attention mechanisms and gradient-based explanations
  • Algorithmic fairness and bias detection
  • Fairness definitions and impossibility results
  • Practical implementation with Python tools

Prerequisites: Machine learning basics, Python
Level: Intermediate to Advanced
Duration: 2-3 weeks


πŸ”· Bayesian Methods & MCMC

Advanced MCMC Methods

πŸ“‚ 03-bayesian-methods/mcmc/

Learning Objectives:

  • Understand Bayesian inference and posterior distributions
  • Derive Metropolis-Hastings acceptance probability
  • Implement MCMC algorithms from scratch
  • Apply Hamiltonian Monte Carlo for efficient sampling
  • Use No-U-Turn Sampler (NUTS) for automatic tuning
  • Diagnose convergence using R-hat and ESS
  • Apply MCMC to real Bayesian models

Topics Covered:

  • Bayesian inference fundamentals
  • Metropolis-Hastings algorithm
  • Hamiltonian Monte Carlo and leapfrog integration
  • No-U-Turn Sampler (NUTS)
  • Convergence diagnostics (trace plots, R-hat, ESS)
  • Applications (Bayesian regression, hierarchical models)

Prerequisites: Probability theory, calculus, Python
Level: Advanced
Duration: 3-4 weeks
Code: Complete Python implementations (8,000+ lines)


Bayesian Machine Learning

πŸ“‚ 03-bayesian-methods/bayesian-machine-learning/

Learning Objectives:

  • Apply Bayesian inference to machine learning problems
  • Build Bayesian linear and logistic regression models
  • Implement Gaussian Processes for regression
  • Understand Bayesian neural networks
  • Perform approximate inference (VI, EP)
  • Apply Bayesian optimization for hyperparameter tuning
  • Quantify predictive uncertainty

Topics Covered:

  • Bayesian linear regression
  • Gaussian Processes
  • Bayesian neural networks
  • Variational inference
  • Bayesian optimization
  • Uncertainty quantification

Prerequisites: Bayesian statistics, machine learning, Python
Level: Advanced
Duration: 3-4 weeks


πŸ”· Causal Inference & Econometrics

Causal Inference

πŸ“‚ 04-causal-inference/causal-inference-fundamentals/

Learning Objectives:

  • Understand potential outcomes framework
  • Draw and interpret causal DAGs
  • Implement Instrumental Variables (IV/2SLS)
  • Apply Regression Discontinuity Design
  • Use Difference-in-Differences methods
  • Estimate propensity scores and perform matching
  • Apply synthetic control methods
  • Identify and address confounding

Topics Covered:

  • Potential outcomes and causal graphs
  • Instrumental Variables and weak instruments
  • Regression Discontinuity (sharp and fuzzy)
  • Difference-in-Differences and event studies
  • Propensity score methods
  • Synthetic controls
  • Modern methods (Callaway-Sant'Anna, Sun-Abraham)

Prerequisites: Statistics, econometrics, R or Python
Level: Advanced
Duration: 4-5 weeks
Code: Python & R implementations (11,000+ lines)


πŸ”· Time Series & Forecasting

Time Series Analysis

πŸ“‚ 05-time-series/time-series-forecasting/

Learning Objectives:

  • Analyze time series components (trend, seasonality)
  • Test for and achieve stationarity
  • Build ARIMA and SARIMA models
  • Implement VAR models for multivariate series
  • Apply state space models and Kalman filter
  • Use LSTM and Transformers for forecasting
  • Evaluate forecasting accuracy
  • Apply hybrid methods (Prophet, N-BEATS)

Topics Covered:

  • Stationarity and unit root tests
  • ARMA, ARIMA, SARIMA models
  • Vector Autoregression (VAR)
  • State space models and Kalman filter
  • Forecasting and evaluation
  • Deep learning for time series (LSTM, GRU)
  • Transformer models (TFT, Autoformer, Informer)
  • Hybrid approaches (ES-RNN, N-BEATS, Prophet)

Prerequisites: Statistics, linear algebra, Python
Level: Intermediate to Advanced
Duration: 3-4 weeks


πŸ”· Optimization & Computational Methods

Optimization for Data Science

πŸ“‚ 01-foundations/optimization/

Learning Objectives:

  • Formulate optimization problems
  • Understand convexity and its implications
  • Derive and apply KKT conditions
  • Implement gradient descent variants
  • Apply momentum and adaptive methods (Adam, AdamW)
  • Solve constrained optimization problems
  • Use evolutionary algorithms for black-box optimization
  • Apply Bayesian optimization for hyperparameter tuning
  • Optimize neural network training

Topics Covered:

  • Convex optimization fundamentals
  • Gradient descent (batch, SGD, mini-batch)
  • Momentum methods and Nesterov acceleration
  • Adaptive learning rates (AdaGrad, RMSProp, Adam)
  • Constrained optimization (Lagrangian, KKT, penalties)
  • Evolutionary algorithms (GA, ES, PSO, CMA-ES)
  • Bayesian optimization
  • Multi-objective optimization

Prerequisites: Calculus, linear algebra, Python
Level: Intermediate to Advanced
Duration: 3-4 weeks


πŸ”· Applied Data Science

A/B Testing & Experimentation

πŸ“‚ 04-causal-inference/ab-testing/

Learning Objectives:

  • Design statistically rigorous A/B tests
  • Calculate required sample sizes
  • Perform hypothesis testing correctly
  • Control for multiple comparisons
  • Understand statistical power and effect sizes
  • Apply sequential testing methods
  • Analyze experimental results
  • Avoid common pitfalls (peeking, p-hacking)

Topics Covered:

  • Experimental design
  • Hypothesis testing and p-values
  • Sample size calculations
  • Multiple testing corrections
  • Bayesian A/B testing
  • Sequential analysis
  • Common pitfalls and best practices

Prerequisites: Statistics, probability
Level: Intermediate
Duration: 1-2 weeks


πŸ—οΈ Repository Structure

academic-presentations/
β”œβ”€β”€ README.md                           # This file
β”œβ”€β”€ CONTRIBUTING.md                     # Contribution guidelines
β”œβ”€β”€ CHANGELOG.md                        # Version history
β”œβ”€β”€ LICENSE                            # CC BY-SA 4.0 for content
β”‚
β”œβ”€β”€ .github/                           # πŸ€– GitHub Actions automation
β”‚   β”œβ”€β”€ workflows/
β”‚   β”‚   β”œβ”€β”€ compile-latex.yml         # Auto-compile PDFs
β”‚   β”‚   β”œβ”€β”€ check-links.yml           # Verify all URLs
β”‚   β”‚   └── generate-previews.yml     # Create PDF previews
β”‚   β”œβ”€β”€ dependabot.yml                # Dependency updates
β”‚   └── markdown-link-check-config.json
β”‚
β”œβ”€β”€ shared/                            # πŸ”„ Shared resources
β”‚   β”œβ”€β”€ theme/                        # 🎨 Professional LaTeX theme
β”‚   β”‚   β”œβ”€β”€ esmad_beamer_theme.sty   # Custom Beamer theme
β”‚   β”‚   β”œβ”€β”€ esmad_beamer_theme_highcontrast.sty
β”‚   β”‚   β”œβ”€β”€ STYLE_GUIDE.md           # Theme documentation
β”‚   β”‚   └── template_presentation.tex # Example template
β”‚   └── bibliographies/               # πŸ“š Reference libraries (140+ papers)
β”‚       β”œβ”€β”€ mcmc_references.bib      # MCMC methods (30+ refs)
β”‚       β”œβ”€β”€ causal_inference_references.bib # Causal inference (50+ refs)
β”‚       └── statistical_learning_references.bib # ML/Stats (60+ refs)
β”‚
β”œβ”€β”€ 00-programming-fundamentals/      # πŸ’» Programming Basics
β”‚   └── r-programming/
β”‚       └── presentation/
β”‚           └── R_programming.tex
β”‚
β”œβ”€β”€ 01-foundations/                   # πŸ“Š Core Foundations
β”‚   β”œβ”€β”€ statistical-modeling/
β”‚   β”‚   └── presentation/            # Statistical Learning Theory
β”‚   β”œβ”€β”€ feature-engineering/
β”‚   β”‚   └── presentation/            # Feature Engineering
β”‚   β”œβ”€β”€ pca/
β”‚   β”‚   └── presentation/            # Principal Component Analysis
β”‚   └── optimization/
β”‚       └── presentation/            # Optimization for Data Science
β”‚
β”œβ”€β”€ 02-deep-learning/                 # 🧠 Deep Learning
β”‚   β”œβ”€β”€ deep-learning-fundamentals/
β”‚   β”‚   └── presentation/            # Deep Learning Fundamentals
β”‚   └── reinforcement-learning/
β”‚       └── presentation/            # Reinforcement Learning
β”‚
β”œβ”€β”€ 03-bayesian-methods/              # 🎲 Bayesian Statistics
β”‚   β”œβ”€β”€ mcmc/
β”‚   β”‚   β”œβ”€β”€ presentation/            # MCMC Methods
β”‚   β”‚   └── exercises/               # MCMC Exercises
β”‚   └── bayesian-machine-learning/
β”‚       └── presentation/            # Bayesian ML
β”‚
β”œβ”€β”€ 04-causal-inference/              # βš–οΈ Causal Methods
β”‚   β”œβ”€β”€ causal-inference-fundamentals/
β”‚   β”‚   β”œβ”€β”€ presentation/            # Causal Inference Fundamentals
β”‚   β”‚   └── exercises/               # Causal Inference Exercises
β”‚   └── ab-testing/
β”‚       └── presentation/            # A/B Testing & Experimentation
β”‚
β”œβ”€β”€ 05-time-series/                   # ⏱️ Time Series
β”‚   └── time-series-forecasting/
β”‚       └── presentation/            # Time Series Analysis
β”‚
β”œβ”€β”€ 06-advanced-topics/               # πŸ”¬ Advanced Topics
β”‚   β”œβ”€β”€ explainable-ai/
β”‚   β”‚   └── presentation/            # Explainable AI
β”‚   └── computer-science/
β”‚       └── presentation/            # OOP & Streaming Pipelines
β”‚
β”œβ”€β”€ 07-capstone-projects/             # πŸŽ“ Projects
β”‚   β”œβ”€β”€ industry-focus/              # Industry applications
β”‚   β”œβ”€β”€ project-guides/              # Project guidelines
β”‚   └── prerequisites/               # Prerequisites
β”‚
└── 08-data-science-applications-course/  # 🎯 Applied Course
    β”œβ”€β”€ presentation/                # Full course materials
    └── assessments/                 # Course assessments

🎨 Professional Theme & Styling

All presentations use the ESMAD Beamer Theme for consistent, professional appearance:

Theme Features

βœ… Professional color palette (ESMAD Blue, accents)
βœ… Custom environments (theorems, definitions, examples, alerts)
βœ… Mathematical notation helpers (\Normal, \E, \Var, etc.)
βœ… Code listing styles with syntax highlighting
βœ… Author information with ORCID integration
βœ… Slide templates (title, TOC, contact, references)

Usage

\documentclass[aspectratio=169]{beamer}
\usepackage{../../../shared/theme/esmad_beamer_theme}

% Author info
\authorname{Your Name}
\authoremail{your.email@university.edu}
\authororcid{0000-0000-0000-0000}

\title{Your Presentation}
\date{\today}

\begin{document}
\begin{frame}
  \titlepage
\end{frame}

% Your content...

\contactslide
\end{document}

See shared/theme/STYLE_GUIDE.md for complete documentation.

πŸ”§ Getting Started

Prerequisites

LaTeX Distribution:

# Ubuntu/Debian
sudo apt-get install texlive-full

# macOS
brew install --cask mactex

# Windows
# Download and install MiKTeX or TeX Live

Python Environment (for code examples):

pip install numpy scipy matplotlib seaborn pandas scikit-learn statsmodels
pip install torch tensorflow  # For deep learning examples
pip install shap lime  # For XAI examples

R Environment (for R examples):

install.packages(c(
  "AER", "rdrobust", "fixest", "did",  # Causal inference
  "caret", "recipes", "mice",           # Feature engineering
  "forecast", "vars", "fable"           # Time series
))

Compiling Presentations

Manual compilation:

cd 02-deep-learning/deep-learning-fundamentals/presentation/
pdflatex deep_learning_beamer.tex
pdflatex deep_learning_beamer.tex  # Run twice for references

Using latexmk (recommended):

cd 02-deep-learning/reinforcement-learning/presentation/
latexmk -pdf rl_beamer.tex

Automated compilation:

  • Push to GitHub β†’ GitHub Actions automatically compiles all PDFs
  • Download compiled PDFs from Actions artifacts or Releases

Running Code Examples

Python:

# MCMC examples (if code/ directory exists with implementations)
# Example references are embedded in presentation materials

# Exercises and assessments
cd 03-bayesian-methods/mcmc/exercises/
pdflatex mcmc_exercises.tex

Exercises:

# MCMC exercises
cd 03-bayesian-methods/mcmc/exercises/
pdflatex mcmc_exercises.tex

# Causal inference exercises
cd 04-causal-inference/causal-inference-fundamentals/exercises/
pdflatex causal_inference_exercises.tex

πŸ“– For Students

Recommended Learning Paths

Path 1: Machine Learning Fundamentals

  1. Statistical Learning (4 weeks)
  2. Feature Engineering (2 weeks)
  3. Optimization (3 weeks)
  4. Explainable AI (2 weeks)

Path 2: Deep Learning Specialization

  1. Deep Learning Fundamentals (4 weeks)
  2. Optimization (focus on neural networks)
  3. Reinforcement Learning (4 weeks)
  4. Time Series Analysis (focus on deep methods)

Path 3: Causal & Bayesian Methods

  1. Causal Inference (5 weeks)
  2. Bayesian ML (4 weeks)
  3. MCMC Methods (3 weeks)
  4. A/B Testing (2 weeks)

Study Tips

  • πŸ“š Start with slides to understand concepts
  • πŸ’» Run code examples to see methods in action
  • πŸ“ Complete exercises to test understanding
  • πŸ“– Read references for deeper knowledge
  • 🀝 Join discussions (create GitHub issues)

πŸ‘¨β€πŸ« For Educators

Course Integration

These materials can be integrated into:

  • Graduate courses in Data Science/Statistics/CS
  • Professional training programs
  • Workshop series
  • Seminar courses

Customization

  1. Fork this repository
  2. Customize presentations for your needs
  3. Add your own examples and exercises
  4. Maintain attribution (CC BY-SA 4.0)

Assessment Resources

Use the materials in assessments/:

  • Quizzes for each topic
  • Midterm and final exams
  • Grading rubrics
  • Project ideas

πŸ”¬ For Researchers

Citation

If you use these materials in your research or teaching, please cite:

@misc{ribeiro2025academic,
  author = {Ribeiro, Diogo},
  title = {Academic Presentations: Comprehensive Data Science Course Materials},
  year = {2025},
  publisher = {GitHub},
  url = {https://github.com/diogoribeiro7/academic-presentations},
  note = {ESMAD \& Mysense.ai}
}

Using the Bibliographies

All presentations reference comprehensive BibTeX files:

\usepackage[backend=bibtex]{biblatex}
\addbibresource{../../../shared/bibliographies/mcmc_references.bib}

% In document
\cite{metropolis1953}
\cite{hoffman2014}

% At end
\printbibliography

Available:

  • shared/bibliographies/mcmc_references.bib: 30+ MCMC papers
  • shared/bibliographies/causal_inference_references.bib: 50+ causal inference papers
  • shared/bibliographies/statistical_learning_references.bib: 60+ ML/stats papers

All include DOIs for easy access.

🀝 Contributing

We welcome contributions! See CONTRIBUTING.md for guidelines.

How to Contribute

  1. Fork the repository
  2. Create a feature branch
  3. Make your changes
  4. Test compilation and code
  5. Submit a pull request

Contribution Types

  • πŸ› Fix errors in presentations
  • πŸ“š Add new presentations
  • πŸ’‘ Improve existing content
  • πŸ“– Enhance documentation
  • πŸ§ͺ Add code examples
  • πŸ“ Create exercises
  • 🎨 Improve theme/styling

πŸ€– Automation & CI/CD

GitHub Actions Workflows

  • compile-latex.yml: Auto-compiles all LaTeX on push
  • check-links.yml: Verifies all URLs and DOIs weekly
  • generate-previews.yml: Creates PDF preview gallery
  • dependabot.yml: Keeps dependencies updated

PDF Preview Gallery

View slide previews at: https://diogoribeiro7.github.io/academic-presentations/

πŸ“Š Repository Statistics

  • πŸ“š 15+ comprehensive presentations
  • πŸ’» 27,000+ lines of code (Python & R)
  • πŸ“– 140+ curated references with DOIs
  • πŸ“ 14 pages of exercises (2 comprehensive problem sets)
  • 🎨 1 professional LaTeX theme with full documentation
  • πŸ€– Fully automated PDF compilation and testing

πŸ“„ License

Content (Presentations & Exercises)

Licensed under Creative Commons Attribution-ShareAlike 4.0 International

You are free to:

  • βœ… Share β€” copy and redistribute
  • βœ… Adapt β€” remix, transform, and build upon

Under the terms:

  • πŸ“ Attribution required
  • πŸ”„ ShareAlike for derivatives

Code

Code examples licensed under MIT License

πŸ“ž Contact & Collaboration

Professional Inquiries

Research Interests

  • Markov Chain Monte Carlo and Bayesian computation
  • Machine learning and deep learning
  • Causal inference and econometrics
  • Financial risk modeling
  • Time series analysis and forecasting

Collaboration Opportunities

  • πŸŽ“ Guest lectures and workshops
  • 🏒 Corporate training programs
  • πŸ”¬ Research collaborations
  • πŸ“ Joint publications
  • 🌐 Conference presentations

🌟 Acknowledgments

  • ESMAD for institutional support
  • Mysense.ai for industry applications and insights
  • Students and colleagues for valuable feedback
  • Open source community for tools and inspiration
  • Academic community for rigorous peer review

πŸ“ˆ Version History

See CHANGELOG.md for detailed version history.


Last Updated: January 2025
Repository Maintainer: Diogo Ribeiro
Status: βœ… Actively maintained
Latest Release: View releases

GitHub stars GitHub forks GitHub watchers

About

Professional-grade presentations on advanced statistics, MCMC methods, and data science applications. Created by ESMAD faculty and Mysense.ai Lead Data Scientist. Includes theoretical foundations, modern algorithms, and industry implementations.

Topics

Resources

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •