What Integrates with Kedro?

Find out what Kedro integrations exist in 2025. Learn what software and services currently integrate with Kedro, and sort them by reviews, cost, features, and more. Below is a list of products that Kedro currently integrates with:

  • 1
    Jupyter Notebook Reviews
    Open-source web application, the Jupyter Notebook, allows you to create and share documents with live code, equations, and visualizations. Data cleaning and transformation, numerical modeling, statistical modeling and data visualization are just a few of the many uses.
  • 2
    Python Reviews
    Definitive functions are the heart of extensible programming. Python supports keyword arguments, mandatory and optional arguments, as well as arbitrary argument lists. It doesn't matter if you are a beginner or an expert programmer, Python is easy to learn. Python is easy to learn, whether you are a beginner or an expert in other languages. These pages can be a helpful starting point to learn Python programming. The community hosts meetups and conferences to share code and much more. The documentation for Python will be helpful and the mailing lists will keep in touch. The Python Package Index (PyPI), hosts thousands of third-party Python modules. Both Python's standard library and the community-contributed modules allow for endless possibilities.
  • 3
    pandas Reviews
    Pandas is an open-source data analysis and manipulation tool that is fast, flexible, flexible, and easy to use. It was built on top the Python programming language. Tools for reading and writing data between memory data structures and various formats: CSV, text files, Microsoft Excel, SQL databases and the fast HDF5 format. Intelligent data alignment and integrated handling missing data: Use a powerful group engine to perform split-apply/combine operations on data sets. Time series-functionality: date range generation and frequency conversion, moving window statistics, date shifting and lagging. You can even create domain-specific offsets and join time sequences without losing data.
  • 4
    Docker Reviews
    Docker eliminates repetitive, tedious configuration tasks and is used throughout development lifecycle for easy, portable, desktop, and cloud application development. Docker's complete end-to-end platform, which includes UIs CLIs, APIs, and security, is designed to work together throughout the entire application delivery cycle. Docker images can be used to quickly create your own applications on Windows or Mac. Create your multi-container application using Docker Compose. Docker can be integrated with your favorite tools in your development pipeline. Docker is compatible with all development tools, including GitHub, CircleCI, and VS Code. To run applications in any environment, package them as portable containers images. Use Docker Trusted Content to get Docker Official Images, images from Docker Verified Publishings, and more.
  • 5
    Plotly Dash Reviews
    Dash & Dash Enterprise allow you to build and deploy analytic web applications using Python, R, or Julia. No JavaScript or DevOps are required. The world's most successful companies offer AI, ML and Python analytics at a fraction of the cost of full-stack development. Dash is the way they do it. Apps and dashboards that run advanced analytics such as NLP, forecasting and computer vision can be delivered. You can work in Python, R, or Julia. Reduce costs by migrating legacy per-seat license software to Dash Enterprise's unlimited end-user pricing model. You can deploy and update Dash apps faster without an IT or DevOps staff. You can create pixel-perfect web apps and dashboards without having to write any CSS. Kubernetes makes it easy to scale. High availability support for mission-critical Python apps
  • 6
    Vertex AI Reviews
    Fully managed ML tools allow you to build, deploy and scale machine-learning (ML) models quickly, for any use case. Vertex AI Workbench is natively integrated with BigQuery Dataproc and Spark. You can use BigQuery to create and execute machine-learning models in BigQuery by using standard SQL queries and spreadsheets or you can export datasets directly from BigQuery into Vertex AI Workbench to run your models there. Vertex Data Labeling can be used to create highly accurate labels for data collection. Vertex AI Agent Builder empowers developers to design and deploy advanced generative AI applications for enterprise use. It supports both no-code and code-driven development, enabling users to create AI agents through natural language prompts or by integrating with frameworks like LangChain and LlamaIndex.
  • 7
    Amazon SageMaker Reviews
    Amazon SageMaker, a fully managed service, provides data scientists and developers with the ability to quickly build, train, deploy, and deploy machine-learning (ML) models. SageMaker takes the hard work out of each step in the machine learning process, making it easier to create high-quality models. Traditional ML development can be complex, costly, and iterative. This is made worse by the lack of integrated tools to support the entire machine learning workflow. It is tedious and error-prone to combine tools and workflows. SageMaker solves the problem by combining all components needed for machine learning into a single toolset. This allows models to be produced faster and with less effort. Amazon SageMaker Studio is a web-based visual interface that allows you to perform all ML development tasks. SageMaker Studio allows you to have complete control over each step and gives you visibility.
  • 8
    Matplotlib Reviews
    Matplotlib is a Python library that allows you to create interactive, animated, or static visualizations. Matplotlib makes difficult things simple and easy. Many third-party packages extend and build upon Matplotlib functionality. These include several higher-level plotting interfaces such as seaborn, HoloViews and ggplot.
  • 9
    Azure Machine Learning Reviews
    Accelerate the entire machine learning lifecycle. Developers and data scientists can have more productive experiences building, training, and deploying machine-learning models faster by empowering them. Accelerate time-to-market and foster collaboration with industry-leading MLOps -DevOps machine learning. Innovate on a trusted platform that is secure and trustworthy, which is designed for responsible ML. Productivity for all levels, code-first and drag and drop designer, and automated machine-learning. Robust MLOps capabilities integrate with existing DevOps processes to help manage the entire ML lifecycle. Responsible ML capabilities – understand models with interpretability, fairness, and protect data with differential privacy, confidential computing, as well as control the ML cycle with datasheets and audit trials. Open-source languages and frameworks supported by the best in class, including MLflow and Kubeflow, ONNX and PyTorch. TensorFlow and Python are also supported.
  • 10
    Apache Spark Reviews

    Apache Spark

    Apache Software Foundation

    Apache Spark™, a unified analytics engine that can handle large-scale data processing, is available. Apache Spark delivers high performance for streaming and batch data. It uses a state of the art DAG scheduler, query optimizer, as well as a physical execution engine. Spark has over 80 high-level operators, making it easy to create parallel apps. You can also use it interactively via the Scala, Python and R SQL shells. Spark powers a number of libraries, including SQL and DataFrames and MLlib for machine-learning, GraphX and Spark Streaming. These libraries can be combined seamlessly in one application. Spark can run on Hadoop, Apache Mesos and Kubernetes. It can also be used standalone or in the cloud. It can access a variety of data sources. Spark can be run in standalone cluster mode on EC2, Hadoop YARN and Mesos. Access data in HDFS and Alluxio.
  • 11
    MLflow Reviews
    MLflow is an open-source platform that manages the ML lifecycle. It includes experimentation, reproducibility and deployment. There is also a central model registry. MLflow currently has four components. Record and query experiments: data, code, config, results. Data science code can be packaged in a format that can be reproduced on any platform. Machine learning models can be deployed in a variety of environments. A central repository can store, annotate and discover models, as well as manage them. The MLflow Tracking component provides an API and UI to log parameters, code versions and metrics. It can also be used to visualize the results later. MLflow Tracking allows you to log and query experiments using Python REST, R API, Java API APIs, and REST. An MLflow Project is a way to package data science code in a reusable, reproducible manner. It is based primarily upon conventions. The Projects component also includes an API and command line tools to run projects.
  • 12
    Dask Reviews
    Dask is free and open-source. It was developed in collaboration with other community projects such as NumPy and pandas. Dask uses existing Python data structures and APIs to make it easy for users to switch between NumPy/pandas and scikit-learn-powered versions. Dask's schedulers can scale to thousands of node clusters, and its algorithms have been tested at some of the most powerful supercomputers around the world. You don't necessarily need a large cluster to get started. Dask ships schedulers that can be used on personal computers. Many people use Dask to scale computations on their laptops, using multiple cores and their disk for extra storage. Dask exposes lower level APIs that allow you to build custom systems for your own applications. This allows open-source leaders to parallelize their own packages, and business leaders to scale custom business logic.
  • 13
    Azure Databricks Reviews
    Azure Databricks allows you to unlock insights from all your data, build artificial intelligence (AI), solutions, and autoscale your Apache Spark™. You can also collaborate on shared projects with other people in an interactive workspace. Azure Databricks supports Python and Scala, R and Java, as well data science frameworks such as TensorFlow, PyTorch and scikit-learn. Azure Databricks offers the latest version of Apache Spark and allows seamless integration with open-source libraries. You can quickly spin up clusters and build in an Apache Spark environment that is fully managed and available worldwide. Clusters can be set up, configured, fine-tuned, and monitored to ensure performance and reliability. To reduce total cost of ownership (TCO), take advantage of autoscaling or auto-termination.
  • 14
    Kubeflow Reviews
    Kubeflow is a project that makes machine learning (ML), workflows on Kubernetes portable, scalable, and easy to deploy. Our goal is not create new services, but to make it easy to deploy the best-of-breed open source systems for ML to different infrastructures. Kubeflow can be run anywhere Kubernetes is running. Kubeflow offers a custom TensorFlow job operator that can be used to train your ML model. Kubeflow's job manager can handle distributed TensorFlow training jobs. You can configure the training controller to use GPUs or CPUs, and to adapt to different cluster sizes. Kubeflow provides services to create and manage interactive Jupyter Notebooks. You can adjust your notebook deployment and compute resources to meet your data science requirements. You can experiment with your workflows locally and then move them to the cloud when you are ready.
  • 15
    Apache Airflow Reviews

    Apache Airflow

    The Apache Software Foundation

    Airflow is a community-created platform that allows programmatically to schedule, author, and monitor workflows. Airflow is modular in architecture and uses a message queue for managing a large number of workers. Airflow can scale to infinity. Airflow pipelines can be defined in Python to allow for dynamic pipeline generation. This allows you to write code that dynamically creates pipelines. You can easily define your own operators, and extend libraries to suit your environment. Airflow pipelines can be both explicit and lean. The Jinja templating engine is used to create parametrization in the core of Airflow pipelines. No more XML or command-line black-magic! You can use standard Python features to create your workflows. This includes date time formats for scheduling, loops to dynamically generate task tasks, and loops for scheduling. This allows you to be flexible when creating your workflows.
  • Previous
  • You're on page 1
  • Next