Make Your Company Data Driven. Connect to any data source, easily visualize, dashboard and share your data.
-
Updated
Oct 11, 2020 - JavaScript
{{ message }}
Make Your Company Data Driven. Connect to any data source, easily visualize, dashboard and share your data.
This is more a question than a feature request.
When parsing JSON files, I need to sanitize the field names so field with spaces becomes field_with_spaces.
I want to preserve the original name as well, metadata about the column if you like :)
There is a metadata field on StructField, but it is internal.
Why is this internal, is it possible or desirable to expose it?
Apache Spark Connector for Azure Cosmos DB
Enabling Continuous Data Processing with Apache Spark and Azure Event Hubs
machine learning for genomic variants
Scalable Data Science, course sets in big data Using Apache Spark over databricks and their mathematical, statistical and computational foundations using SageMath.
Tools for Deploying Databricks Solutions in Azure
Testing framework for Databricks notebooks
code, labs and lectures for the course
MLFLow Tracking Server based on Docker and AWS S3
Collection of Sample Databricks Spark Notebooks ( mostly for Azure Databricks )
Collection of Machine Learning Examples for Azure Databricks
(117th place - Top 26%) Deep learning using Keras and Spark for the "Store Item Demand Forecasting" Kaggle competition.
Terraform Databricks provider
A simplified, autogenerated API client interface using the databricks-cli package
Connect your Spark Databricks clusters Log4J output to the Application Insights Appender
Connect-Databricks.ps1 uses "https://login.microsoftonline.com" as part of the URI to connect. When retrieving a token for a non-AzureCloud tenant (e.g. AzureUSGovernment) the URI root would be different (e.g., "https://login.microsoftonline.us"). As such, cannot use this task to deploy to other tenant types. Would be helpful to be able to specify an Azure Environment and connect to the right e
A solution for on-demand training and serving of Machine Learning models, using Azure Databricks and MLflow
A working example of DevOps & operationalisation applied to Machine Learning and AI
OctopuFS library helps managing cloud storage, ADLSgen2 specifically. It allows you to operate on files (moving, copying, setting ACLs) in very efficient manner. Designed to work on databricks, but should work on any other platform as well.
Using U-Net Model to Detect Wildfire from Satellite Imagery
Spark stream consumption commit in kafka consumer group
Are you like me , a Senior Data Scientist, wanting to learn more about how to approach DevOps, specifically when you using Databricks (workspaces, notebooks, libraries etc) ? Set up using @Azure @databricks
Terraform script to deploy almost all Azure Data Services
A collection of “cookbook-style” scripts for simplifying data engineering and machine learning in Apache Spark.
Add a description, image, and links to the databricks topic page so that developers can more easily learn about it.
To associate your repository with the databricks topic, visit your repo's landing page and select "manage topics."
I have a simple regression task (using a LightGBMRegressor) where I want to penalize negative predictions more than positive ones. Is there a way to achieve this with the default regression LightGBM objectives (see https://lightgbm.readthedocs.io/en/latest/Parameters.html)? If not, is it somehow possible to define (many example for default LightGBM model) and pass a custom regression objective?