diponkar.paul

Diponkar Paul has 14 years of experience in the IT industry and specialized in the Business Intelligence domain, currently working as a Manager of Data Engineering at OMERS, Canada. Throughout his career, he designs and developed medium and large data warehouses and delivered business-critical analytical solutions. He was leading the Toronto PASS Group; which is part of the world's largest data professional community named PASS and Lead of Toronto Data Professionals Community. He got a diverse background and has been working and studying in different Geo locations including Canada, Denmark, the UK, and Sweden. He writes a technical blog and speaks, he loves to share his learning within the community.
  • Interests: Data, Travel, Swimming, Cycling, cross-country skate
  • Blog: http://allaboutdata.ca/
  • Jobs: Senior Data Engineer
  • Skills: Big data, Business Intelligence, Database, ETL, ELT, Data Warehouse

SQLServerCentral Article

Handling Row-level Errors in ADF Data Flows

If you are working with ADF (Azure Data Factory) data flows, then you may have noticed there was a new feature released in November 2020, which is useful to capture any error while inserting/updating the records in a SQL database. This article will describe how to setup the error row handling feature and why it's […]

(1)

You rated this post out of 5. Change rating

2021-06-16 (first published: )

5,666 reads

SQLServerCentral Article

Step by Step Guide to Install Jupyter Notebook

Whether you work as a Data Engineer or a Data Scientist, a Jupyter Notebook is a helpful tool. One of the projects I was working required a comparison of two parquet files. This is mainly a schema comparison, not a data comparison. Though the two .parquet were created from two different sources, the outcome should […]

You rated this post out of 5. Change rating

2021-05-17

5,504 reads

SQLServerCentral Article

Dynamically Add a Timestamp To Files in Azure Data Factory

This article will describe how to add your local timestamp at the end of the each file in Azure Data Factory (ADF). In general, ADF gets a UTC timestamp, so we need to convert the timestamp from UTC to EST, since our local time zone is EST. For example, if the input Source file name […]

(2)

You rated this post out of 5. Change rating

2021-04-22

30,235 reads

SQLServerCentral Article

How to simulate Case statement in Azure Data Factory (ADF) compared with SSIS?

This post will describe how you use a CASE statement in Azure Data Factory (ADF). If you are coming from SSIS background, you know a piece of SQL statement will do the task. However let's see how do it in SSIS and the very same thing can be achieved in ADF. Problem statement For my […]

(2)

You rated this post out of 5. Change rating

2020-11-12

22,477 reads

Blogs

AI: Blog a Day – Day 5: Notebooks, Hugging face models and Fine Tuning

By

Continuing from Day 4 where we learned Encoder, Decoder, and Attention Mechanism, today we...

AI: Blog a Day – Day 4: Transformers – Encoder, Decoder, and Attention

By

Continuing from Day 3 where we covered LLM models open/closed and their parameters, Today...

Flyway Tips: Multiple Projects

By

One of the nice things about Flyway Desktop is that it helps you manage...

Read the latest Blogs

Forums

Can an Azure App Service Managed Identity be used for SQL Login?

By jasona.work

I'm fairly certain I know the answer to this from digging into it yesterday,...

Azure Synapse database refresh

By Sreevathsa Mandli

Hi Team, I am trying to refresh the Azure Synapse Dedicated pool from production...

how to write this query?

By water490

hi everyone I am not sure how to write the query that will produce...

Visit the forum

Question of the Day

Fun with JSON I

I have some data in a table:

CREATE TABLE #test_data
(
    id INT PRIMARY KEY,
    name VARCHAR(100),
    birth_date DATE
);

-- Step 2: Insert rows  
INSERT INTO #test_data
VALUES
(1, 'Olivia', '2025-01-05'),
(2, 'Emma', '2025-03-02'),
(3, 'Liam', '2025-11-15'),
(4, 'Noah', '2025-12-22');
If I run this query, how many rows are returned?
SELECT *
FROM OPENJSON(
     (
         SELECT t.* FROM #test_data AS t FOR JSON PATH
     )
             ) t;

See possible answers