...importantes clientes. ⬇⬇ Requisitos ⬇⬇ Características de los profesionales especialistas del Servicio: - Sólidos conocimientos en herramientas FICO-Blaze/RMA y DMP Streaming, - Sólidos conocimientos en Arquitectura y Sistemas TI. - Validación de Pruebas de Concepto (POC). - Experiencia en Arquitectura de Integración. - Conocimiento en Blaze-RMA, DMPS, Hbase-Intermediate, Hadoop Intermediate, Hive Intermediate y Kafka. Solución de FICO DMPS y Blaze contemplan las siguientes actividades tanto correctivas como también evolutivas: · Resolución de dudas sobre herramientas FICO y otras herramientas del proyecto. · Control de acceso en las herramientas de creac...
...el Director técnico para colaborar en la dirección técnica y contribuir a soluciones de problemas que se planteen en este campo ● Realizar revisiones de código Debe tener: ● Fuertes habilidades de programación, idealmente en Python, Scala, Java o un lenguaje similar ● Experiencia en integración continua utilizando software de control de versiones GIT ● Conocimiento en Base de Datos SQL y noSQL (HBase, Mongo, Cassandra,...) ● Conocimientos de sistemas Linux/Unix a nivel profesional ● Persona con gran capacidad de trabajar autonomamente y mucha proactividad Deseable: ● Conocimientos con framework .NET utilizando lenguajes de programación como C#, C++, Visual Basic .NET, Python, Prolog ● Conocimientos de Sistemas de Cloud (Azure, AW...
Se trata de un proyecto generación de estrategia de Big Data para Business Analytics , las fases que queremos cubrir son las de revisión y auditoría de las fuentes de datos (estructurados - no estructurados) y el apoyo técnico en l...operativos y redes. • Arquitectura de centro de procesos de datos y creación de Data Lakes (Cloudera,HortonWorks,MapR) • Familiarización con entornos de procesamientos modernos a escala masiva (Big Data) y/o en tiempo real: Hadoop/Mapreduce, HBase Scala/Spark, dataflow, Storm, Flume. • Conocimientos del entorno Salesforce •...
...poder modelar la mejor arquitectura posible para un determinado problema concreto de negocio. Buscamos: Titulados en Informática, Matemáticas, Estadística, etc Requisitos mínimos Experiencia con Git Lenguajes de programación: Java, Python, Scala, R... Experiencia en desarrollo de procesos MapReduce en Hadoop, Spark o Flink Manipulación de datos en diferentes DB Nosql como Cassandra, Mongo o HBase Requisitos deseados Experiencia en desarrollo de procesos Real Time con Storm, Spark o Flink Experiencia o conocimiento de las herramientas del ecosistema Hadoop Librerías y/o tecnologías de manipulación y visualización de datos: Tableau, QlikView, Pentaho... Capacidad de trabajo en equipo con metodolog&iacut...
Sopra selecciona, para importante proyectos de la Compañía 4 CONSULTORES BIG DATA con acreditada experiencia de al menos 2 años en:- Explotación, analítica y manejo de los datos almacenados.- Linux (administración general y networking).- Programación Java.- Flume, logstash o syslog (nivel básico).- Instalación y uso de Hadoop (básico). - Apache Kafka. - BBDD NoSQL (Mongo, Cassandra, HBase, etc)Se valorará conocimientos y/o experiencia con: - Hortonworks o Cloudera. - Python. - Storm o Spark. - Scala. - ElasticSearch. - Splunk. - Hive / profesionales con capacidad de trabajo en equipo, ganas de aprender, mejorar y asumir nuevos retos, iniciativa y responsabilidad en el trabajo Ofrecemos: - Estabilidad labo...
Experiencia con Hortonworks,Conocimientos de Sandbox tipo: HBase; Hive; Spark; Oozie; Gestión con Ambari;...Experiencia en proyectos de implantaciónDeseable que haya trabajado con QlicView a nivel de instalación.-
Sopra selecciona, para importante proyectos de la Compañía 4 CONSULTORES BIG DATA con acreditada experiencia de al menos 2 años en:- Explotación, analítica y manejo de los datos almacenados.- Linux (administración general y networking).- Programación Java.- Flume, logstash o syslog (nivel básico).- Instalación y uso de Hadoop (básico). - Apache Kafka. - BBDD NoSQL (Mongo, Cassandra, HBase, etc)Se valorará conocimientos y/o experiencia con: - Hortonworks o Cloudera. - Python. - Storm o Spark. - Scala. - ElasticSearch. - Splunk. - Hive / profesionales con capacidad de trabajo en equipo, ganas de aprender, mejorar y asumir nuevos retos, iniciativa y responsabilidad en el trabajo Ofrecemos: - Estabilidad labo...
Sopra selecciona, para importante proyectos de la Compañía 4 CONSULTORES BIG DATA con acreditada experiencia de al menos 2 años en:- Explotación, analítica y manejo de los datos almacenados.- Linux (administración general y networking).- Programación Java.- Flume, logstash o syslog (nivel básico).- Instalación y uso de Hadoop (básico). - Apache Kafka. - BBDD NoSQL (Mongo, Cassandra, HBase, etc)Se valorará conocimientos y/o experiencia con: - Hortonworks o Cloudera. - Python. - Storm o Spark. - Scala. - ElasticSearch. - Splunk. - Hive / profesionales con capacidad de trabajo en equipo, ganas de aprender, mejorar y asumir nuevos retos, iniciativa y responsabilidad en el trabajo Ofrecemos: - Estabilidad labo...
En Vates buscamos una persona con experiencia en administración y operaciones de soluciones deployadas en el cloud, particularmente en AWS y muy buen manejo de Linux (especialmente Ubuntu y CentOS) para sumerse a equipo de :- Experiencia en administración y operación del ecosistema de Apache en lo referido a big data: HDFS, ZooKeeper, HBase, Kafka.- Conocimientos de OpenTSDB.- Experiencia trabajando con AWS: Networking, security, Load Balancer, AMIs, RDS (conocimientos de SQL y administración de MySQL), SES.- Experiencia trabajando con REST APIs, node.js, javascript, Apache Tomcat.- Muy buen manejo de entornos Linux, scripting, troubleshooting y analisys de :- Conocimientos administrando Cloudera- Conocimie...
En Vates buscamos una persona con experiencia en administración y operaciones de soluciones deployadas en el cloud, particularmente en AWS y muy buen manejo de Linux (especialmente Ubuntu y CentOS) para sumerse a equipo de :- Experiencia en administración y operación del ecosistema de Apache en lo referido a big data: HDFS, ZooKeeper, HBase, Kafka.- Conocimientos de OpenTSDB.- Experiencia trabajando con AWS: Networking, security, Load Balancer, AMIs, RDS (conocimientos de SQL y administración de MySQL), SES.- Experiencia trabajando con REST APIs, node.js, javascript, Apache Tomcat.- Muy buen manejo de entornos Linux, scripting, troubleshooting y analisys de :- Conocimientos administrando Cloudera- Conocimie...
BluePatagon empres...en Big Data (Hadoop - Hortonworks), para importante cliente de CABA. Experiencia: mínima de 1 año en tecnologías Hadoop: Experiencia en desarrollo de aplicaciones (MapReduce + HDFS). Big Data: Familiaridad con el ecosistema (Hive, Pig, HBase, etc). y conceptos de escalabilidad, análisis en tiempo real, procesamiento de datos distribuidos. Linux: Uso Avanzado (manejo de servicios de SO, administración, shell scripting, seguridad) Programacion: POO (Java preferentemente, Python) Base de datos: RDBMS (Oracle, MySQL, PostgreSQL), NoSQL (HBase, Cassandra) Data Exchange y configuración: JSON + XML Deseable Data Science Research, Data Product Design, Data Visualization, Machine Learning, Data Mining...
...administración de Windows Server. La experiencia en administración de cuentas de usuario, gestión de paquetes, endurecimiento de la seguridad, y ajuste de rendimiento. Comprensión de la supervisión del sistema y soluciones de registro de aplicación (Nagios, grafito, Splunk / Logstash). Experiencia con lenguajes de scripting (Per, Python, golpe, Powershell ). Conocimientos de SQL Server y NoSQL (MongoDB, HBase) para manejar grandes conjuntos de datos de la aplicación. Experiencia en configuración de herramientas de automatización (Títeres, Chef, SCCM ) Inglés Avanzado. Requisitos deseables: Conocimiento de ASP.NET MVC, C #, Java u otro código abierto Comprensión del Cloud...
Vacantes Data4 Python + Django Web Developers Data4 esta buscando desarrolladores en Django/Python para el departamento de Data Driven Web Platforms. El departamento esta encargado de elaborar complejas plataformas usando Django, PostgreSQL+ PostGIS / Mysql, Memcached, Haystack con Solr, y su integración a Big Data con Hadoop, Hbase, R y las librerías en Python para machine learning como PyBrain, NumPy, SciPY. Trabajamos bajo la metodología ágil SCRUM con mucha apertura a la experimentación y el auto aprendizaje. Nuestros desarrolladores trabajan desde la arquitectura de la base de datos hasta la construcción de excelentes UI y UX de las plataformas. Data4 tiene su base en la Ciudad de México. Responsabilidades<...
... <strong>Habilidades requeridas:</strong> 6 años (Senior) de experiencia en la construcción y gestión de calidad de producción, los productos /soluciones complejas de software. Nivel de mando expertas en Java y C++ con experiencia en programación. Nivel experto en diseño de software orientado a objetos y servicios. Gran entendimiento en Java Server-side Gran experiencia en NoSQL /HBASE, Cassandra o similares, Neo4j). Voluntad de aprender e implementar nuevas tecnologías BigData, según sea necesario. Iniciativa y capacidad de trabajar de manera independiente y en equipo. Experiencia con Sotrm en soluciones real time analytics. Experiencia en procesamiento paralelo (MPI, OpenMP) como ventaja compe...
Hiring: DevOps Developers DevOps Engineer Requirements: -Expertise in CentOS, Ubuntu, Debian -Experience with HBase, Hadoop, Storm, ArangoDB, Prometheus -Strong benchmarking and system scalability testing skills Responsibilities: -Configure and optimize high-availability servers and databases -Implement monitoring and performance evaluation tools -Review and enhance DevOps workflows Long-term opportunities available!
My Ubuntu server already has Docker running both of these repos: • • I now need a quick, clean setup of the supporting services so everything talks to each other: 1. Spin up HBase (Docker is available) and create the tables required by archival-rpc and ingestor-rpc. 2. Spin up Kafka, link it to HBase, and confirm messages flow from ingestor-rpc into the HBase tables. 3. Prove the full chain works by querying archival-rpc and showing live data being served. Docker is already installed and running; you may add docker-compose or individual containers—whichever is fastest. A concise README or set of commands that reproduces your setup on a fresh box will complete the job. I’m looking for
...data from multiple sources: Logs, transactions, social media, IoT devices, sensors, clickstreams, etc. Tools: Apache Flume, Kafka, Sqoop (for importing from databases). --- 2. Data Storage Big Data needs distributed, fault-tolerant storage (not just normal databases). Options: HDFS (Hadoop Distributed File System) – stores data across many machines. NoSQL Databases – MongoDB, Cassandra, HBase. Cloud Storage – AWS S3, Google Cloud Storage, Azure Data Lake. --- 3. Data Processing Once stored, data must be processed (batch or real-time). Batch Processing (large chunks at once): Hadoop MapReduce Apache Spark (faster, in-memory processing) Stream Processing (real-time, continuous): Apache Kafka + Spark Streaming Apache Flink / Storm --- ...
...learning model, optimize large-scale data pipelines, or query massive datasets — I can deliver solutions tailored to your needs. What I Can Do for You: AI & Machine Learning Deep Learning (CNN, LSTM, Transformer) Image & Text Processing (NLP, OCR, Captioning) Model training, tuning, and evaluation TensorFlow, PyTorch, scikit-learn Big Data Engineering ETL pipelines with Apache Spark Hive, HBase, Hadoop ecosystem Data warehouse optimization Stream & batch data processing Databases SQL (PostgreSQL, MySQL) NoSQL (MongoDB, Neo4j, Cassandra) Graph databases & complex queries Development Tools Python scripting & automation REST APIs, FastAPI, Flask Git, Docker, Linux, Jupyter...
I have a CSV file containing sales data, with a size between 1-10 GB. ...CSV file containing sales data, with a size between 1-10 GB. I need a skilled data engineer to ingest this data into RDS, clean it, and load it into HBase using Apache Sqoop. Ultimately, the cleaned data will be analyzed using MapReduce on HBase. Key tasks: - Ingest CSV data into RDS - Clean the data - Load cleaned data into HBase using Apache Sqoop - Conduct analysis using MapReduce on HBase The data cleaning process will involve: - Removing duplicates - Handling missing values - Fixing formatting inconsistencies The ideal freelancer for this project should have: - Proficiency in data engineering and management - Experience with RDS, HBase, Apache Sqoop and MapReduce - Strong ski...
I need someone to handle the processing and analysis of my sales team performance data. The data is currently stored in CSV files an...performance data. The data is currently stored in CSV files and I need it loaded into Amazon RDS, specifically a MySQL instance. Tasks include: - Loading the CSV data into Amazon RDS - Cleaning the data by removing duplicates, handling missing/null values and standardizing formats - Loading the cleaned data into HBase using Apache Sqoop - Performing analysis using MapReduce Ideal skills for this project are: - Proficient in MySQL - Experienced in data cleaning and processing - Familiar with HBase and Apache Sqoop - Competent in using MapReduce for data analysis I am looking for a professional who can deliver high-quality work and has a kee...
I'm looking for a professional who can help me transfer relational data from my Amazon RDS MySQL database to HBase using Apache Sqoop. This is a one-time project, so I'm looking for someone who can complete this task efficiently and effectively. Ideal skills for this job include: - Proficiency in Apache Sqoop - Experience with MySQL databases - Knowledge of HBase - Ability to handle relational data Please provide evidence of your skills in your bid. Thank you.
I have a sales data CSV file that needs to be ingested into AWS RDP. The data will need to be cleaned (including removing duplicates, handling missing values, and normalizing the data) before it is loaded into HBase using Apache Sqoop. Finally, I need some analysis to be performed on the data using Map Reduce. Ideal Skills and Experience: - Proficiency in AWS and RDP - Experience with data cleaning and processing - Familiarity with HBase and Apache Sqoop - Strong understanding of Map Reduce - Past experience with sales data is a plus Please include in your application a brief description of your relevant experience.
I need assistance with managing a product dataset contained in a CSV file. The tasks include: - Uploading the CSV dataset into AWS RDS using a provided schema. - Data cleaning, which involves removing null values and duplicate entries. - Transferring the cleaned data to HBase using Apache Sqoop. - Conducting trend analysis on the dataset using MapReduce. Ideal candidates for this project should have strong experience with AWS RDS, data cleaning, Apache Sqoop, HBase, and MapReduce. Please note that no additional data cleaning is required beyond the mentioned tasks.
I'm looking for a skilled data engineer to assist with my dataset. Key Tasks: - Upload and structure a CSV dataset (1GB to 10GB) in AWS RDS - Move the data into HBase using Apache Sqoop - Clean the data, which involves handling missing values, removing duplicates, and standardizing formats - Use MapReduce to process and analyze sales trends, store and vendor performance, and category breakdowns - Deliver clear insights to guide sales, marketing, and vendor strategies Ideal Skills and Experience: - Proficiency in AWS services, specifically RDS - Experience with Apache Sqoop and HBase - Strong data cleaning and preparation skills - Familiarity with MapReduce for data analysis - Ability to interpret data and deliver actionable insights
Data Setup & Migration Import the liquor sales data into AWS RDS, using a provided data dictionary to guide the schema. If needed, split large files into smaller parts to keep things efficient. Use Apache Sqoop to move the data from RDS to HBase, and design a clean, scalable schema for HBase. Make sure the data is accurate and consistent in both systems. Cleaning Things Up Get rid of missing, incomplete, or broken records. Standardize formats across fields (especially dates, categories, etc.). Deduplicate entries to keep the data tidy and usable. Batch Processing & Analysis Use MapReduce to process and analyze data in bulk. Key insights we’re looking for: Revenue breakdown by store, county, and liquor category. Top performers – best-selling categ...
...(2012–2020). Below is a breakdown of the work involved: 1. Data Ingestion & Preparation Import the "Liquor Sales" dataset into AWS RDS, setting up the schema based on a provided data dictionary. If needed, split large files into manageable chunks to optimize performance. Use Apache Sqoop to migrate the data from RDS to HBase, designing a suitable schema with appropriate column families and row keys. Ensure data consistency and validate the integrity of the data in both RDS and HBase. 2. Data Cleaning Clean the dataset to improve quality and ensure accurate analysis: Remove incomplete or missing data. Fix formatting issues and invalid entries. Standardize categorical data and normalize field formats. Deduplicate records to maintain clean dataset...
...liquor sales dataset spanning from 2012 to 2020. This involves uploading the data into AWS RDS, transferring it to HBase, and ensuring its integrity before analysis. Key Tasks: - Data Ingestion: Upload the dataset into AWS RDS, defining an appropriate schema based on the provided data dictionary. Utilize Apache Sqoop to transfer data from AWS RDS to HBase, creating a tall schema with suitable column families and row keys. - Data Integrity: Validate the integrity of the ingested data in both systems. - Data Cleaning: Remove incomplete values, handle inconsistent data, standardise categories, eliminate duplicates. Ideal Candidate: - Proficient in using AWS RDS, Apache Sqoop, and HBase. - Experienced in data cleaning and preparation for analysis. - Detail-oriented fo...
...for querying S3 data AWS EMR (Hadoop/Spark) – for big data processing AWS Step Functions – for workflow orchestration AWS Data Pipeline – for scheduling data workflows AWS CloudWatch – for monitoring AWS CloudFormation/Terraform – for infrastructure as code IAM Roles & Policies – for security and access control Data Storage Management-MySQL, MongoDB, SQL Server, Cassandra, Snowflake, PostgreSQL, HBase, Teradata, Hive DevOps Infrastructure: Jenkins, Azure DevOps, Docker, Kubernetes, Terraform, Ansible Apache Spark & Kafka Python (Pandas, NumPy) for scripting and transformation Tableau / Power BI for visualization If you’re someone with solid AWS Data Engineering experience and love teaching others practically, I’d love to le...
I'm in need of an experienced AWS EMR & Hadoop Administrator. Your primary focus will be managing EMR clusters, optimizing big data workflows Key Responsibilities: - Setting up and maintaining EMR clusters - Performance monitoring and troubleshooting issues - Constructing an application server based on provided instructions - Strong expertise required in Hive, Spark, HBase, IAM security, and cost optimization Skills & Experience: - Proficient with Unix as the technology stack for the application server - Prior experience building application servers is essential - Familiarity with automation tools like Terraform, AWS CLI, or CloudFormation is a plus - Freelance/Remote opportunity – Please apply with your relevant experience.
...and optimize servers running CentOS, Ubuntu, and Debian. -Work with distributed systems, including HBase, Hadoop, and Storm. -Configure and maintain a high-availability ArangoDB cluster with redundancy features. -Conduct comprehensive benchmarking and high-availability testing to evaluate system performance and scalability under various conditions. -Set up and integrate monitoring interfaces such as Prometheus to ensure system health and performance monitoring. -Independently review and test the effectiveness of DevOps work and provide constructive feedback. Required Skills & Expertise: -Strong knowledge of server configuration for CentOS, Ubuntu, and Debian. -Hands-on experience with HBase, Hadoop, and Storm. -Proficiency in configuring and managing high-availabilit...
#Your code goes here import '' import '' def jbytes(*args) { |arg| arg.to_s.to_java_bytes } end def put_many(table_name, row, column_values) table = (@, table_name) p = (*jbytes(row)) do |column, value| family, qualifier = (':') (jbytes(family, qualifier), jbytes(value)) end (p) end # Call put_many function with sample data put_many 'wiki', 'DevOps', { "text:" => "What DevOps IaC do you use?", "revision:author" => "Frayad Gebrehana", "revision:comment" => "Terraform" } # Get data from the 'wiki' table get 'wiki', 'DevOps' #Do not remove the exit call below exit
...Visualization of JanusGraph with Elasticsearch Integration for Relationship Analysis in Banking" Requirements Analysis: a. Conduct stakeholder interviews to gather system requirements b. Document use cases and user stories c. Define data schema and relationship mapping for JanusGraph d. Assess technical constraints and system integrations Planning and Design: a. Select the datastore (HBase or Cassandra) after analysing performance and scalability b. Define the JanusGraph schema, data model, and query patterns c. Plan data migration strategy and sequence from Elasticsearch to JanusGraph d. Design the algorithm for relationship creation between Main party and Other party e. Evaluate visualization libraries and choose the most appropriate for the Link Analysis cha...
...Visualization of JanusGraph with Elasticsearch Integration for Relationship Analysis in Banking" Requirements Analysis: a. Conduct stakeholder interviews to gather system requirements b. Document use cases and user stories c. Define data schema and relationship mapping for JanusGraph d. Assess technical constraints and system integrations Planning and Design: a. Select the datastore (HBase or Cassandra) after analysing performance and scalability b. Define the JanusGraph schema, data model, and query patterns c. Plan data migration strategy and sequence from Elasticsearch to JanusGraph d. Design the algorithm for relationship creation between Main party and Other party e. Evaluate visualization libraries and choose the most appropriate for the Link Analysis cha...
...looking for an advanced Hadoop trainer for an online training program. I have some specific topics to be covered as part of the program, and it is essential that the trainer can provide in-depth knowledge and expertise in Hadoop. The topics to be discussed include Big Data technologies, Hadoop administration, Data warehousing, MapReduce, HDFS Architecture, Cluster Management, Real Time Processing, HBase, Apache Sqoop, and Flume. Of course, the trainer should also have good working knowledge about other Big Data topics and techniques. In addition to the topics mentioned, the successful candidate must also demonstrate the ability to tailor the course to meet the learner’s individual needs, making sure that the classes are engaging and fun. The trainer must also possess out...
I am looking for a freelancer who c...through WebEx meetings. Here are the project requirements: Specific Azure topics: - Azure Networking Assistance type: - Virtual Assistance Preferred meeting type: - WebEx Meeting and AZURE Azure Data Factory (ADE), Azure DataBricks, Azure Data Lake Services (ADLS), Azure Blob Services, Azure SQL DB, Azure Active Directory (AAD), Azure Dev Ops. Languages: Scala, Core Java, Python Databases Hive, Hbase Data Ingestion: Sqoop, Kafka, Spark Streaming Data Visualization:Table and AZURE:ADF Databricks Azure Skills and experience: - Strong understanding of Azure Networking - Experience in providing virtual assistance - Proficiency in conducting WebEx meetings If you have the required skills and experience, please bid on this summary
...SupportLocation: Austin, TXDuration: 12 Months Job Description: We are looking someone having strong experience in production support, administration and Development experience with Hadoop technologies.• Minimum Experience 8 Years• Must have Hands-on experience on managing Multiple Hortonworks Clusters. Troubleshooting, Maintaining and Monitoring is the key responsibility here.• Must be conversant in HBase, OpenBSD & Grafana related issues handling in order to ensure the data flow is smooth and consistent.• Experience on Kafka for stream processing of Data• Experience in Deployment of new services, patching of hosts etc.• Good hands-on experience on Linux (preferably Red hat) server platform• Should have knowledge in at least one o...
...topics like AWS Azure GCP DigitalOcean Heroku Alibaba Linux Unix Windows Server (Active Directory) MySQL PostgreSQL SQL Server Oracle MongoDB Apache Cassandra Couchbase Neo4J DynamoDB Amazon Redshift Azure Synapse Google BigQuery Snowflake SQL Data Modelling ETL tools (Informatica, SSIS, Talend, Azure Data Factory, etc.) Data Pipelines Hadoop framework services (e.g. HDFS, Sqoop, Pig, Hive, Impala, Hbase, Flume, Zookeeper, etc.) Spark (EMR, Databricks etc.) Tableau PowerBI Artificial Intelligence Machine Learning Natural Language Processing Python C++ C# Java Ruby Golang Node.js JavaScript .NET Swift Android Shell scripting Powershell HTML5 AngularJS ReactJS VueJS Django Flask Git CI/CD (Jenkins, Bamboo, TeamCity, Octopus Deploy) Puppet/Ansible/Chef Docker Kubernetes ECS/EKS Test...
.../ Define the problem. Create Tables with constraints Design a Schema based on tables and explain the schema. Create primary keys, foreign keys. Create Procedures. Create functions. Create Views Create Index Use of the following Clauses: Example : order by, between, group by, having, order by, AND, OR, with Use Aggregate Functions Use of nested queries, Scalar Subquery. Part 2 has to be done in HBASE Create Tables – 4 tables with Column family and columns Column family - 5 column families: Make sure have different parameter. Ex: versions Minimum 4 Columns in each Column family Insert records Delete records Perform basic queries like your assignment1 Try to extract data using timestamp Insert partial data in a row Describe table. Check table status – enabled or disable...
.../ Define the problem. Create Tables with constraints Design a Schema based on tables and explain the schema. Create primary keys, foreign keys. Create Procedures. Create functions. Create Views Create Index Use of the following Clauses: Example : order by, between, group by, having, order by, AND, OR, with Use Aggregate Functions Use of nested queries, Scalar Subquery. Part 2 has to be done in HBASE Create Tables – 4 tables with Column family and columns Column family - 5 column families: Make sure have different parameter. Ex: versions Minimum 4 Columns in each Column family Insert records Delete records Perform basic queries like your assignment1 Try to extract data using timestamp Insert partial data in a row Describe table. Check table status – enabled or disable...
Looking for Flutter ( Dart) developer work for long term project with fixed monthly payment . Required skills : 1: At least 2 app published 2: Dart 3: API 4: Sqlite ,hbase 5: In app purchase integration experience 6: Bluetooth experience
...oriented discussion. Must Have: ● At least 6+ years of total IT experience ● At least 4+ years of experience in design and development using Hadoop technology stack and programming languages ● Hands-on experience in 2 or more areas: o Hadoop, HDFS, MR o Spark Streaming, Spark SQL, Spark ML o Kafka/Flume. o Apache NiFi o Worked with Hortonworks Data Platform o Hive / Pig / Sqoop o NoSQL Databases HBase/Cassandra/Neo4j/MongoDB o Visualisation & Reporting frameworks like D3.js, Zeppellin, Grafana, Kibana Tableau, Pentaho o Scrapy for crawling websites o Good to have knowledge of Elastic Search o Good to have understanding of Google Analytics data streaming. o Data security (Kerberos/Open LDAP/Knox/Ranger) ● Should have a very good overview of the current landscape and ability to...
Data Engineers 6+ yrs : At least 6+ years of total IT experience ● At least 4+ years of experience in design and development using Hadoop technology stack and programming languages ● Hands-on experience in 2 or more areas: o Hadoop, HDFS, MR o Spark Streaming, Spark SQL, Spark ML o Kafka/Flume. o Apache NiFi Worked with Hortonworks Data Platform o Hive / Pig / Sqoop o NoSQL Databases HBase/Cassandra/Neo4j/MongoDB o Visualisation & Reporting frameworks like D3.js, Zeppellin, Grafana, Kibana Tableau, Pentaho o Scrapy for crawling websites o Good to have knowledge of Elastic Search o Good to have understanding of Google Analytics data streaming. o Data security (Kerberos/Open LDAP/Knox/Ranger) ● Should have a very good overview of the current landscape and ability t...
LDAP service on Oracle Linux 7.3 with rpm packages. LDAP for HDP-2.5.3.0 3. Install and configure Ranger HDP service security on Hbase and Solr tables and collection and validate the security with 5 users
...taking advantage of the CI/CD pipelines when possible - Help with troubleshooting and configuration fine-tuning on several platforms (apache, haddoop, hbase etc) - Build and maintain a local testing environment replica for developers. - Help plan for "non hyper cloud" deployments. OpenStack, ProxMox, Kubernetes. All are on the table but the most "appropriate" one must be selected considering the architecture and CI/CD capabilities. - Build and maintain "on prem" alternatives of the AWS structure. This will include hardware planing (server) but also deployment of several VMs (or containers at some point) with techs including php+nginx, hadoop with hbase (and phoenix), sql database (probably mysql) and CEPH object storage. - Be the technical cha...
The purpose of this project is to develop a working prototype of a network monitoring and reporting Platform that receives network health and status, traffic data from several network infrastructure monitoring sources, and produces an aggr...Platform that receives network health and status, traffic data from several network infrastructure monitoring sources, and produces an aggregate of network status data for processing by a data analytics engine. This prototype will be known as NetWatch. The NetWatch solution will utilize data processing and analytics services via the Hadoop infrastructure, and data reporting features of the Hbase or MYSQL/Datameer tool. The prototype will be used by the Network A&E team to determine its viability as a working engine for network status ...
Please have a look at the below stack. 1. Bash Scripting. 2. Hive 3. Scala Spark 4. HBase and other regular big data technologies.
- Backup HBase database on internal infrastructure
We are looking for a machine learning engineer who must have the following experience: 1. python coding: +7 years of experience 2. Machine Leaning: +5 years of experience (Scikit-Learn, TensorFlow, Caffe, MXNet, Keras, XGBoost) 3. AI/Deep Learning: +5 years of experience 4. Cloud computing: AWS, S3, EC2, EMR, SageMaker, ECS, Lambda, IAM 5. distributed computing technology: Hadoop, Spark, HBase, Hive / Impala, or any similar technology Should be an independent developer, NO CONSULTING COMPANY There will be series of technical interview about python coding, machine learning, AI , cloud computing. Candidate must have an excellent skill in python coding and be able to answer challenging python questions during the interview
Design, code, test Hive, Sqoop, HBase, Yarn, UNIX Shell scripting Spark and Scala mandatory You should have working experience in previous projects not a beginner level projects so please be ready to design develop and fix the bugs. Working hours and all We can decide over the chat.
am trying to run hbase backup command and got below error root@machine:~/hbase-2.4.12# hbase backup Error: Could not find or load main class backup Caused by: : backup need to fix that some tips below : Hbase install below Just enable the configuration on xml file and start the hbase and confirm is working well run Hbase on linux Ubuntu some helps below: