A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | T | U | V | W | X | Y | Z | AA | AB | AC | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | |||||||||||||||||||||||||||||
2 | Email us: instructor.success@datacamp.com | Apply Below: | |||||||||||||||||||||||||||
3 | Title | Description | Content Area | Content Type | Technology | ||||||||||||||||||||||||
4 | Distributed AI Model Training in Python | To train models on consumer-grade hardware, you’ll need to optimize the model training process to get the most out of your processing power. This coding course will cover how to optimize the model training process depending on the hardware available. As well as optimizing hardware, the course should cover smart architecture and training loop choices to optimize training loop memory usage and runtimes, such as gradient accumulation, local SGD, low-precision training, and others. Learners aren’t expected to know how different hardware options (e.g., GPU) enable distributed computing, so some computation fundamentals will need to be covered. Applicants are encouraged to use Hugging Face’s accelerate library in this course for accelerating PyTorch model training, but we will accept applications using other libraries with good reasons. | Artificial Intelligence | Course | Python | Application Form | |||||||||||||||||||||||
5 | Deploying AI into Production with FastAPI | Deploying models as APIs has become a popular method for making them easily accessible to systems in a developer-friendly way. This course will cover how to take models from scripts to production using FastAPI to create the API and uvicorn to deploy the server, as well as running the application from a Docker container. As the course progresses, we would like learners to add features to their API, such as authentication and usage limits to replicate many real-world use cases. Prior to the course, learners will be familiar with GET and POST operations in FastAPI. | Artificial Intelligence | Course | Python | Application Form | |||||||||||||||||||||||
6 | Introduction to PySpark | This two- or three-chapter course will teach learners what Spark is and how to use Spark from Python! It will also introduce learners to the challenges of working with big data. Learners will learn to interactively work with DataFrames and PySpark SQL to process and manipulate large datasets. The course should replace the existing Introduction to PySpark course and be an entry point to our PySpark curriculum for data and AI engineers, data scientists, and machine learning engineers. The prerequisite will be Joining Data with pandas. | Emerging Curriculum | Course | PySpark | Application Form | |||||||||||||||||||||||
7 | Advanced Git | This course explores advanced topics and techniques using Git. Learners will discover why and how to perform various (non-default) merges, including squashing and rebasing. The course will introduce concepts such as cherry-picking, bisecting, and reference logging, concluding with how to work with submodules and worktrees. Before taking the course, learners will be familiar with saving, reverting, and searching in Git, as well as working with branches and remotes. | Emerging Curriculum | Course | Git | Application Form | |||||||||||||||||||||||
8 | Case Study: Databricks | This case study will provide learners hands-on experience using Databricks to tackle real-world problems. Using a fictitious company within a specific industry (such as finance, FMCG, or healthcare), learners will take on the role of a Data Analyst. Learners will take on tasks associated with managing data, utilizing SQL in Databricks, creating visuals and dashboards, and performing analytical processes. The learner works his or her way through the analytics problem from A-Z by receiving guidance in the instructions. This interactive case study will bring together the learners' Databricks knowledge picked up during the early stages of their journey in learning the tool and help learners apply the skills needed to pass the Databricks Certified Data Analyst Associate exam. | Business Intelligence | Course | Databricks | Application Form | |||||||||||||||||||||||
9 | Data Transformation in KNIME | This course delves into the powerful world of data transformation using KNIME. Learners will start with the basics, gaining a solid understanding of KNIME’s essential data transformation techniques. The course will dive into practical applications, teaching how to effectively manipulate, clean, and integrate data within KNIME. They will also discover how to clean data by removing duplicates, handling missing values, applying filters, and integrating data from various sources. Additionally, learners will explore the Math Formula node to perform complex calculations and create new variables based on mathematical expressions. By the end, learners will be proficient in importing data, utilizing various transformation nodes, and building workflows to optimize data processes. | Business Intelligence | Course | KNIME | Application Form | |||||||||||||||||||||||
10 | Data Manipulation in KNIME | Help learners unlock the power of data manipulation with KNIME! This course will teach the essential techniques for merging, aggregating, and exporting data using KNIME’s robust tools. Learners will discover how to concatenate multiple tables, perform value lookups, and execute complex join operations. It will allow them to master group by and pivot operations with the Row Aggregator and understand the GroupBy, Pivot, and unpivoting nodes. It will also teach learners to write data in CSV, Excel, and other formats, handle remote file systems, and explore utility nodes for file management and database writing. By the end of this course, learners will be proficient in using KNIME to streamline data processes, enhance analysis capabilities, and efficiently handle data manipulation tasks | Business Intelligence | Course | KNIME | Application Form | |||||||||||||||||||||||
12 | Transform and Analyze Data with Microsoft Fabric | This course focuses on implementing data cleaning and preparation processes in Microsoft Fabric. Learners will explore star schemas, normalization, data transformations, and performance optimization. This course will be part of our track covering the Fabric Analytics Engineer Associate certificate from Microsoft (DP-600) and will include interactive exercises in the Fabric environment. For this course, we will assume learners are comfortable ingesting data into Fabric; this course should be about using the data now that it has been ingested. | Cloud | Course | Microsoft Fabric | Application Form | |||||||||||||||||||||||
13 | Plan and Implement a Data Analytics Environment with Microsoft Fabric | This course focuses on the administrative side of Microsoft Fabric. Learners will configure a Fabric environment for a larger team using the admin portal, implement access control for Fabric items, manage workspaces, warehouses, lakehouses, and other administrative Fabric jobs. This course will be part of our track covering the Fabric Analytics Engineer Associate certificate from Microsoft (DP-600) and will include interactive exercises in the Fabric environment. This course will come near the end of our Fabric track; as a result, learners will have a basic understanding of Fabric and its use cases. | Cloud | Course | Microsoft Fabric | Application Form | |||||||||||||||||||||||
14 | Project: Find trends in data | In this project, you will create visualizations and then interpret them to uncover possible trends and other insights from the data. Prerequisites: Intermediate Data Visualization with ggplot2 Libraries/packages: ggplot2 | Data Science | Project | R | Application Form | |||||||||||||||||||||||
15 | |||||||||||||||||||||||||||||
16 | |||||||||||||||||||||||||||||
17 | |||||||||||||||||||||||||||||
18 | |||||||||||||||||||||||||||||
19 | |||||||||||||||||||||||||||||
20 | |||||||||||||||||||||||||||||
21 | NOTE: Courses are color-coded by "Technology" | ||||||||||||||||||||||||||||
22 | Conceptual | ||||||||||||||||||||||||||||
23 | Excel | ||||||||||||||||||||||||||||
24 | Databricks | ||||||||||||||||||||||||||||
25 | Power BI/Tableau/Alteryx/KNIME | ||||||||||||||||||||||||||||
26 | Python | ||||||||||||||||||||||||||||
27 | R | ||||||||||||||||||||||||||||
28 | Snowflake | ||||||||||||||||||||||||||||
29 | SQL | ||||||||||||||||||||||||||||
30 | Cloud | ||||||||||||||||||||||||||||
31 | Java | ||||||||||||||||||||||||||||
32 | PySpark | ||||||||||||||||||||||||||||
33 | Kubernetes | ||||||||||||||||||||||||||||
34 | Hugging Face | ||||||||||||||||||||||||||||
35 | Microsoft Fabric | ||||||||||||||||||||||||||||
36 | Git | ||||||||||||||||||||||||||||
37 | |||||||||||||||||||||||||||||
38 | |||||||||||||||||||||||||||||
39 | |||||||||||||||||||||||||||||
40 | |||||||||||||||||||||||||||||
41 | |||||||||||||||||||||||||||||
42 | |||||||||||||||||||||||||||||
43 | |||||||||||||||||||||||||||||
44 | |||||||||||||||||||||||||||||
45 | |||||||||||||||||||||||||||||
46 | |||||||||||||||||||||||||||||
47 | |||||||||||||||||||||||||||||
48 | |||||||||||||||||||||||||||||
49 | |||||||||||||||||||||||||||||
50 | |||||||||||||||||||||||||||||
51 | |||||||||||||||||||||||||||||
52 | |||||||||||||||||||||||||||||
53 | |||||||||||||||||||||||||||||
54 | |||||||||||||||||||||||||||||
55 | |||||||||||||||||||||||||||||
56 | |||||||||||||||||||||||||||||
57 | |||||||||||||||||||||||||||||
58 | |||||||||||||||||||||||||||||
59 | |||||||||||||||||||||||||||||
60 | |||||||||||||||||||||||||||||
61 | |||||||||||||||||||||||||||||
62 | |||||||||||||||||||||||||||||
63 | |||||||||||||||||||||||||||||
64 | |||||||||||||||||||||||||||||
65 | |||||||||||||||||||||||||||||
66 | |||||||||||||||||||||||||||||
67 | |||||||||||||||||||||||||||||
68 | |||||||||||||||||||||||||||||
69 | |||||||||||||||||||||||||||||
70 | |||||||||||||||||||||||||||||
71 | |||||||||||||||||||||||||||||
72 | |||||||||||||||||||||||||||||
73 | |||||||||||||||||||||||||||||
74 | |||||||||||||||||||||||||||||
75 | |||||||||||||||||||||||||||||
76 | |||||||||||||||||||||||||||||
77 | |||||||||||||||||||||||||||||
78 | |||||||||||||||||||||||||||||
79 | |||||||||||||||||||||||||||||
80 | |||||||||||||||||||||||||||||
81 | |||||||||||||||||||||||||||||
82 | |||||||||||||||||||||||||||||
83 | |||||||||||||||||||||||||||||
84 | |||||||||||||||||||||||||||||
85 | |||||||||||||||||||||||||||||
86 | |||||||||||||||||||||||||||||
87 | |||||||||||||||||||||||||||||
88 | |||||||||||||||||||||||||||||
89 | |||||||||||||||||||||||||||||
90 | |||||||||||||||||||||||||||||
91 | |||||||||||||||||||||||||||||
92 | |||||||||||||||||||||||||||||
93 | |||||||||||||||||||||||||||||
94 | |||||||||||||||||||||||||||||
95 | |||||||||||||||||||||||||||||
96 | |||||||||||||||||||||||||||||
97 | |||||||||||||||||||||||||||||
98 | |||||||||||||||||||||||||||||
99 | |||||||||||||||||||||||||||||
100 | |||||||||||||||||||||||||||||
101 |