Activity
Mon
Wed
Fri
Sun
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
What is this?
Less
More

Memberships

Fabric Dojo ็ป‡็‰ฉ

339 members โ€ข $447/y

Learn Microsoft Fabric

12.7k members โ€ข Free

3 contributions to Learn Microsoft Fabric
How to ingest rest api response in table in lakehouse
Since I have not much experience with API requests I don't know how to proceed further. What I am trying to do is to use PySpark Notebooks for getting first of all the response and then convert it correctly into a dataframe. This needs to work well on scale, because I need to ingest approximately 3 million rows. I can perform simple requests with the request python library, but don't know how I can translate that to creating a solution for big data. Therefore I need to perform paging/looping. The API I am using does only support paging through offset and limit parameters. So I need to loop till all items are retrieved while I still need to ensure that it's not causing overhead and running parallel. However, the output is nested which doesn't make it easier for me. I have issues with loosing data while converting it into a dataframe due to objects hold other objects or a new array with objects and somewhere in between the schema doesn't properly convert. Below I do have an example of all the levels in my json output. "company": { "companyId": "932xxx5stest", "companyCode": "TEST", "_links": [ { "rel": "self", "href": "https://api.test.com/v1/companies/932xxx5stest" } ] } The Copy Data Activity in a pipeline doesn't really work for me because my api doesn't provide really well the count of total items. So I can't extract that by using a pipeline which means i have to do things manually. That's why I prefer the notebook. Any ideas/usefull resources or (your) best practices are welcome! Thanks in advance. If you need more information, please ask and i'll provide some more context.
1 like โ€ข Jun '24
@Maurice Weststrate yeh no probs if you get stuck I will try and help if I can
1 like โ€ข Jul '24
@Maurice Weststrate I've not attempted to do the requests in parallel since the API's I have used are limited by the number of requests they allow per minute and some of them I've had to throttle because one notebook looping was generating too many requests. I've used azure data factory to get REST data but it usually depends on the pagination used and if its simple to get working in data factory or not, I've found notebooks give you extra control. For bookYear filtering, I usually have a Date Dimension set up somewhere in my system so if I need to cycle through years/months or days I'd usually query this to give me what I need i.e. select distinct years from DimDate where date is less than today.
Get value from notebook and pass to next pipeline activity
I feel like this should be easier than it seems. Iโ€™m trying to build an api call in a fabric pipeline. I have a lake house with a delta table that has a column of ID fields. Id like to iterate over each one and use the value to build an api call (one api call for each value). Iโ€™m struggling with outputting the value to the next activity, Iโ€™m trying mssparkutils and lookup activity but I canโ€™t seem to get it. Any ideas?
1 like โ€ข Jun '24
@Will Needham thanks for the work around, just started working with pipelines (been mainly python so far) and was very surprised to find the query option missing from the lookup - this solved my problem cheers
๐Ÿ‘‹ New joiner? Welcome! Start here ๐Ÿ‘‡
Welcome to all new members, here's some links and information to help you get started! ๐—ค๐˜‚๐—ถ๐—ฐ๐—ธ ๐—Ÿ๐—ถ๐—ป๐—ธ๐˜€ ๐˜๐—ผ ๐—ด๐—ฒ๐˜ ๐˜€๐˜๐—ฎ๐—ฟ๐˜๐—ฒ๐—ฑ - For an introduction to this community โ†’ Explore the Welcome Pack - New-ish to Fabric? โ†’ Check out our Fabric Foundation module - Studying for the DP-600? โ†’ Check out the DP-600 Module and the DP-600 category - Studying for the DP-700? โ†’ Check out the DP-700 Module and the DP-700 category - Want to get hands-on? โ†’ Check out Fabric Dojo ๐—›๐—ผ๐˜„ ๐˜๐—ผ ๐—ฒ๐—ป๐—ด๐—ฎ๐—ด๐—ฒ ๐˜„๐—ถ๐˜๐—ต ๐˜๐—ต๐—ฒ ๐—ฐ๐—ผ๐—บ๐—บ๐˜‚๐—ป๐—ถ๐˜๐˜†? - Share your knowledge and experience! Even if you're relatively new to Fabric, or the community, your opinion and experiences are valued here! A great way to earn your first point(s) is to introduce yourself in the thread below ๐Ÿ‘‡๐Ÿ˜€ Thank you for engaging and joining us on this exciting learning journey! ๐Ÿ™ Will
4 likes โ€ข Jun '24
Hi Everyone, I'm Steve from the UK and work as a senior data engineer at Acumine who specialise in data solutions for the construction industry. 20+ years working with data, originally Oracle, then Azure and recently MS Fabric for about 10 months. Interested to keep up to date with the new functionality being added to fabric all the time and learn best practices.
1-3 of 3
Steve Foster
2
11points to level up
@steve-foster-4783
Data Engineer with over 20 years experience started with Oracle then moved to Azure and now mainly Fabric

Online now
Joined May 31, 2024
Powered by