Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Getting Started with Talend Open Studio for Data Integration

You're reading from   Getting Started with Talend Open Studio for Data Integration This is the complete course for anybody who wants to get to grips with Talend Open Studio for Data Integration. From the basics of transferring data to complex integration processes, it will give you a head start.

Arrow left icon
Product type Paperback
Published in Nov 2012
Publisher Packt
ISBN-13 9781849514729
Length 320 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
Jonathan Bowen Jonathan Bowen
Author Profile Icon Jonathan Bowen
Jonathan Bowen
Arrow right icon
View More author details
Toc

Table of Contents (22) Chapters Close

Getting Started with Talend Open Studio for Data Integration
Credits
Foreword
Foreword
About the Author
Acknowledgement
About the Reviewers
www.PacktPub.com
Preface
1. Knowing Talend Open Studio FREE CHAPTER 2. Working with Talend Open Studio 3. Transforming Files 4. Working with Databases 5. Filtering, Sorting, and Other Processing Techniques 6. Managing Files 7. Job Orchestration 8. Managing Jobs 9. Global Variables and Contexts 10. Worked Examples Installing Sample Jobs and Data Resources Index

Modifying data in a database


It is not unusual for data integrations to be simply about getting data from one system, modifying it, then passing it onto another system to consume the classic Extract, Transform and Load (ETL) scenario. However, sometimes we will also need to modify the data in the database that we are sourcing data from or, indeed, the database we are sending data to.

To illustrate this, imagine we have a database table containing customer orders. We need to extract the data out of this table and send it to another system so that the orders can be fulfilled. However, most of the time we will need to be selective about the data that we send over. We cannot select all rows from the table, because presumably some will have already been sent. We could filter our result set by date/time so that the job only runs once per hour and each time it sends data created in the last hour. This would probably work perfectly well for most scenarios, but even this kind of filter is not quite...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime