santhosh0000000
As a Data Engineer, i can design and build robust data pipelines and architectures. Specializing in technologies such as Hadoop, Apache Kafka, Spark
Pinned Repositories
ETL_SAP-HDFS
The Java program uses Apache Spark to connect to a SAP HANA database, retrieve data from a specific table, and then write this data to a Hadoop Distributed File System (HDFS) in CSV format. Here is a brief summary of the changes:
BACKUP---DB
mysql-postgresql-MongoDB-MariaDB
SAP-Hive
Java program that uses Apache Spark to connect to a SAP HANA database, retrieve data from a specific table, and then write this data to a Hive table. This is a common pattern in big data processing pipelines
SAP_Postgreql_ETL
This Python script connects to a SAP HANA database, retrieves data from a specified table, and then transfers this data to a PostgreSQL database.
Boto--S3-MIG
The Script migrate empty file name data from a Scality S3 cluster to a Ceph S3 cluster using the boto library in Python
JDBC_SAP
The Java code is designed to connect to an SAP HANA database, execute a SQL query, and write the results to a CSV file using the OpenCSV library.
ETL_sap_hive
This Java program leverages the Apache Spark framework to read data from a SAP HANA database and save it to a Hive table in ORC format
MYSQL--backup
Config files for my GitHub profile.
postgresql-database
backup postgresql database
sap-hana__HiveDB
The code appears to read configurations from a CSV file (test.csv) and, for each configuration, attempts to read data from an SAP table and then write that data to a Hive table.
santhosh0000000's Repositories
santhosh0000000/crossref_API
This Python script is designed to retrieve data from an external API's, process it, and store it in a PostgreSQL database.
santhosh0000000/-BigDataApp
santhosh0000000/incremental-sap_hive
Python function fetch data that incrementally fetches data from a SAP HANA database using PySpark and stores the data in an ORC format in a Hive table.
santhosh0000000/santhosh0000000
santhosh0000000/ETL_sap_hive
This Java program leverages the Apache Spark framework to read data from a SAP HANA database and save it to a Hive table in ORC format
santhosh0000000/sap-hana__HiveDB
The code appears to read configurations from a CSV file (test.csv) and, for each configuration, attempts to read data from an SAP table and then write that data to a Hive table.
santhosh0000000/Spark_JDBC_parallelize
This Java code contains two classes, App2 and App, which are used to read data from a SAP database into a Spark DataFrame and then write it to an HDFS (Hadoop Distributed File System) location in ORC (Optimized Row Columnar) format.
santhosh0000000/SAP_HIVE_SPARK
The code is designed to read data from an SAP HANA database and store it in a Hive table using PySpark.
santhosh0000000/SAP-HDFS-ORC
This code is to read data from a table (PRODELEMENTS) in a HANA database and then save that data to the Hadoop Distributed File System (HDFS) in ORC format.
santhosh0000000/ETL_SAP-HDFS
The Java program uses Apache Spark to connect to a SAP HANA database, retrieve data from a specific table, and then write this data to a Hadoop Distributed File System (HDFS) in CSV format. Here is a brief summary of the changes:
santhosh0000000/JDBC_SAP
The Java code is designed to connect to an SAP HANA database, execute a SQL query, and write the results to a CSV file using the OpenCSV library.
santhosh0000000/SAP-Hive
Java program that uses Apache Spark to connect to a SAP HANA database, retrieve data from a specific table, and then write this data to a Hive table. This is a common pattern in big data processing pipelines
santhosh0000000/Boto--S3-MIG
The Script migrate empty file name data from a Scality S3 cluster to a Ceph S3 cluster using the boto library in Python
santhosh0000000/postgresql-database
backup postgresql database
santhosh0000000/MYSQL--backup
Config files for my GitHub profile.
santhosh0000000/BACKUP---DB
mysql-postgresql-MongoDB-MariaDB
santhosh0000000/SAP_Postgreql_ETL
This Python script connects to a SAP HANA database, retrieves data from a specified table, and then transfers this data to a PostgreSQL database.