11+ Kofax Jobs in Bangalore (Bengaluru) | Kofax Job openings in Bangalore (Bengaluru)
Apply to 11+ Kofax Jobs in Bangalore (Bengaluru) on CutShort.io. Explore the latest Kofax Job opportunities across top companies like Google, Amazon & Adobe.
A global business process management company
B1 – Data Scientist - Kofax Accredited Developers
Requirement – 3
Mandatory –
- Accreditation of Kofax KTA / KTM
- Experience in Kofax Total Agility Development – 2-3 years minimum
- Ability to develop and translate functional requirements to design
- Experience in requirement gathering, analysis, development, testing, documentation, version control, SDLC, Implementation and process orchestration
- Experience in Kofax Customization, writing Custom Workflow Agents, Custom Modules, Release Scripts
- Application development using Kofax and KTM modules
- Good/Advance understanding of Machine Learning /NLP/ Statistics
- Exposure to or understanding of RPA/OCR/Cognitive Capture tools like Appian/UI Path/Automation Anywhere etc
- Excellent communication skills and collaborative attitude
- Work with multiple teams and stakeholders within like Analytics, RPA, Technology and Project management teams
- Good understanding of compliance, data governance and risk control processes
Total Experience – 7-10 Years in BPO/KPO/ ITES/BFSI/Retail/Travel/Utilities/Service Industry
Good to have
- Previous experience of working on Agile & Hybrid delivery environment
- Knowledge of VB.Net, C#( C-Sharp ), SQL Server , Web services
Qualification -
- Masters in Statistics/Mathematics/Economics/Econometrics Or BE/B-Tech, MCA or MBA
Conversational AI- Product Development company
Chatbot Developer
We are a Conversational AI- Product Development company which is located in the USA, Bangalore.
We are looking for a Senior Chatbot /Javascript Developer to join the Avaamo PSG(delivery) team.
Responsibilities:
- Independent team member for analyzing requirements, designing, coding, and implementing Conversation AI products.
- Its a product expert work closely with IT Managers and Business Groups to gather requirements and translate those into the required technical solution.
- Drive solution implementation using the Conversational design approach.
- Develop, deploy and maintain customized extensions to the Avaamo platform-specific to customer requirements.
- Conduct training and technical guidance sessions for partner and customer development teams.
- Evaluating reported defects and the correction of prioritized defects.
- Travel onsite to customer locations for close support.
- Document how to's and implement best practices for Avaamo product solutions.
Requirements:
- Strong programming experience in javascript, HTML/CSS.
- Experience of creating and consuming REST APIs and SOAP services.
- Strong knowledge and awareness of Web Technologies and current web trends.
- Working knowledge of Security in Web applications and services.
- experience in using the NodeJS framework with good understanding of the underlying architecture.
- Experience of deploying web applications on Linux servers in production environment.
- Excellent communication skills.
Good to haves:
- Full stack experience UI and UX design experience or insights
- Working knowledge of AI, ML and NLP.
- Experience of enterprise systems integration like MS Dynamics CRM, Salesforce, ServiceNow, MS Active Directory etc.
- Experience of building Single Sign On in web/mobile applications.
- Ability to learn latest technologies and handle small engineering teams.
- Bring in industry best practices around creating and maintaining robust data pipelines for complex data projects with/without AI component
- programmatically ingesting data from several static and real-time sources (incl. web scraping)
- rendering results through dynamic interfaces incl. web / mobile / dashboard with the ability to log usage and granular user feedbacks
- performance tuning and optimal implementation of complex Python scripts (using SPARK), SQL (using stored procedures, HIVE), and NoSQL queries in a production environment
- Industrialize ML / DL solutions and deploy and manage production services; proactively handle data issues arising on live apps
- Perform ETL on large and complex datasets for AI applications - work closely with data scientists on performance optimization of large-scale ML/DL model training
- Build data tools to facilitate fast data cleaning and statistical analysis
- Ensure data architecture is secure and compliant
- Resolve issues escalated from Business and Functional areas on data quality, accuracy, and availability
- Work closely with APAC CDO and coordinate with a fully decentralized team across different locations in APAC and global HQ (Paris).
You should be
- Expert in structured and unstructured data in traditional and Big data environments – Oracle / SQLserver, MongoDB, Hive / Pig, BigQuery, and Spark
- Have excellent knowledge of Python programming both in traditional and distributed models (PySpark)
- Expert in shell scripting and writing schedulers
- Hands-on experience with Cloud - deploying complex data solutions in hybrid cloud / on-premise environment both for data extraction/storage and computation
- Hands-on experience in deploying production apps using large volumes of data with state-of-the-art technologies like Dockers, Kubernetes, and Kafka
- Strong knowledge of data security best practices
- 5+ years experience in a data engineering role
- Science / Engineering graduate from a Tier-1 university in the country
- And most importantly, you must be a passionate coder who really cares about building apps that can help people do things better, smarter, and faster even when they sleep
Purpose of Job:
Responsible to lead a team of analysts to build and deploy predictive models to infuse core
business functions with deep analytical insights. The Senior Data Scientist will also work
closely with the Kinara management team to investigate strategically important business questions.
Job Responsibilities:
Lead a team through the entire analytical and machine learning model life cycle:
Define the problem statement
Build and clean datasets
Exploratory data analysis
Feature engineering
Apply ML algorithms and assess the performance
Code for deployment
Code testing and troubleshooting
Communicate Analysis to Stakeholders
Manage Data Analysts and Data Scientists
Qualifications:
Education: MS/MTech/Btech graduates or equivalent with a focus on data science and
quantitative fields (CS, Engineering, Mathematics, Economics)
Work Experience: 5+ years in a professional role with 3+ years in ML/AI
Other Requirements: ⮚ Domain knowledge in Financial Services is a big plus
Skills & Competencies
Technical Skills
⮚ Aptitude in Math and Stats
⮚ Proven experience in the use of Python, SQL, DevOps
⮚ Excellent in programming (Python), stats tools, and SQL
⮚ Working knowledge of tools and utilities - AWS, Git, Selenium, Postman,Prefect, Airflow, PySpark
Soft Skills
⮚ Deep Curiosity and Humility
⮚ Strong communications verbal and written
Machine Learning & Deep Learning – Strong
Experienced in TensorFlow, PyTorch, ONNX, Object Detection, Pretrained Models like YOLO, SSD, Faster RCNN, etc…
Python – Strong
NumPy, Pandas, OpenCV
Problem Solving - strong
C++ - average
It will be good if candidate have working experience in C++ in any domain
Note :: Looking for Immediate to 30 days of Notice Period
Expert in Machine Learning (ML) & Natural Language Processing (NLP).
Expert in Python, Pytorch and Data Structures.
Experience in ML model life cycle (Data preparation, Model training and Testing and ML Ops).
Strong experience in NLP, NLU and NLU using transformers & deep learning.
Experience in federated learning is a plus
Experience with knowledge graphs and ontology.
Responsible for developing, enhancing, modifying, optimizing and/or maintaining applications, pipelines and codebase in order to enhance the overall solution.
Experience working with scalable, highly-interactive, high-performance systems/projects (ML).
Design, code, test, debug and document programs as well as support activities for the corporate systems architecture.
Working closely with business partners in defining requirements for ML applications and advancements of solution.
Engage in specifications in creating comprehensive technical documents.
Experience / Knowledge in designing enterprise grade system architecture for solving complex problems with a sound understanding of object-oriented programming and Design Patterns.
Experience in Test Driven Development & Agile methodologies.
Good communication skills - client facing environment.
Hunger for learning, self-starter with a drive to technically mentor cohort of developers. 16. Good to have working experience in Knowledge Graph based ML products development; and AWS/GCP based ML services.
Our Client company is into Telecommunications. (SY1)
- Participate in full machine learning Lifecycle including data collection, cleaning, preprocessing to training models, and deploying them to Production.
- Discover data sources, get access to them, ingest them, clean them up, and make them “machine learning ready”.
- Work with data scientists to create and refine features from the underlying data and build pipelines to train and deploy models.
- Partner with data scientists to understand and implement machine learning algorithms.
- Support A/B tests, gather data, perform analysis, draw conclusions on the impact of your models.
- Work cross-functionally with product managers, data scientists, and product engineers, and communicate results to peers and leaders.
- Mentor junior team members
Who we have in mind:
- Graduate in Computer Science or related field, or equivalent practical experience.
- 4+ years of experience in software engineering with 2+ years of direct experience in the machine learning field.
- Proficiency with SQL, Python, Spark, and basic libraries such as Scikit-learn, NumPy, Pandas.
- Familiarity with deep learning frameworks such as TensorFlow or Keras
- Experience with Computer Vision (OpenCV), NLP frameworks (NLTK, SpaCY, BERT).
- Basic knowledge of machine learning techniques (i.e. classification, regression, and clustering).
- Understand machine learning principles (training, validation, etc.)
- Strong hands-on knowledge of data query and data processing tools (i.e. SQL)
- Software engineering fundamentals: version control systems (i.e. Git, Github) and workflows, and ability to write production-ready code.
- Experience deploying highly scalable software supporting millions or more users
- Experience building applications on cloud (AWS or Azure)
- Experience working in scrum teams with Agile tools like JIRA
- Strong oral and written communication skills. Ability to explain complex concepts and technical material to non-technical users
- Use data to develop machine learning models that optimize decision making in Credit Risk, Fraud, Marketing, and Operations
- Implement data pipelines, new features, and algorithms that are critical to our production models
- Create scalable strategies to deploy and execute your models
- Write well designed, testable, efficient code
- Identify valuable data sources and automate collection processes.
- Undertake to preprocess of structured and unstructured data.
- Analyze large amounts of information to discover trends and patterns.
Requirements:
- 2+ years of experience in applied data science or engineering with a focus on machine learning
- Python expertise with good knowledge of machine learning libraries, tools, techniques, and frameworks (e.g. pandas, sklearn, xgboost, lightgbm, logistic regression, random forest classifier, gradient boosting regressor, etc)
- strong quantitative and programming skills with a product-driven sensibility
- Strong in problem solving, algorithms and data structures
- Proficient in Python
- Hands on experience in technologies and tools related to any of NLP, Deep learning, Machine learning, Conversational AI
- Experience with knowledge Graph or any graph based system is plus
- Able to train and deploy models
- Broad knowledge of machine learning algorithms and principles
- Performance profiling and Tuning
- Communicate and propose solutions to business challenges
- Familiarity with at least one of the cloud computing infrastructure - GCP/AWS
- Keep abreast with the latest technological advances
- Team mentoring and leadership skills.