Arun kumar Jegarkal

(806) 620-8017 · arunjeg@gmail.com

I am a Programmer/Developer with 6+ years of extensive experience in Python, Amazon Web Services, Microsoft Azure, Spark, Tableau, Power BI, Data Analytics and ETL technologies; worked for various clients and projects in every phase of Software Development Life Cycle, from business requirement gathering to project delivery. Excellent design and integration problem solving skills.

During these years, I got chance to interact with individuals and teams across the globe and provide service to the best of my knowledge. Over the years I have gained the skillsets of adapting and developing creative applications, websites, Reports, Dashboards, Azure pipelines, Informatica workflows, maintaining of customer relations, ability to take ownership of tasks, applying proper problem-solving strategies, proper documentation and mindset to think out of box solutions. All these duties were accomplished following customer requirements.

Apart from my work experience I have been fortunate to pursue my master’s degree in Computer Science, where I was exposed to latest tools and techniques along with protocols of industry implementation. With clear understanding of industry working along with education of latest tools and techniques to solve business and data problems. I find myself highly motivated and confident to deliver great service.

Experience

Sr Data Engineer
Data Engineer

Peterbilt Motors Company (Paccar Inc.), Denton, TX

Key Responsibilities:
Data Engineering focuses on making possible fast, accurate, and reliable access to data. I build data pipelines, manage a data warehouse, and support the production use of our data. I advocate for good data practices and make sure that our business users are able to make good data driven decisions.

Primary Duties:
• Provide engineering on modern, cloud-based data processing technology stack
• Build data pipelines, data validation frameworks, job schedules with emphasis on automation and scale
• Contribute to overall architecture, framework, and design patterns to store and process high data volumes
• Ensure product and technical features are delivered to spec and on-time
• Design and implement features in collaboration with product owners, reporting analysts / data analysts, and business partners within an Agile / Scrum methodology
• Proactively support product health by building solutions that are automated, scalable, and sustainable be relentlessly focused on minimizing defects and technical debt
• Focus on making possible fast, accurate, and reliable access to data.
• Interact with Peterbilt engineers, analysts and data scientists to gather the requirements.
• Create the build requirement document by translating the end users requirements into developable features.
• Build the flowchart, architecture designs of the code pipelines based on the requirements using the lucid charts.
• Based on the architecture design, build pipeline in Amazon Web Services (AWS) Cloud in combination of Python, Java and Shell Scripting.
• Unit and Integration Test case document is create for the Services to be implemented.
• Perform the Unit testing of each service implemented in AWS cloud based on the test cases document.
• Perform Integration testing to test the entire pipeline of services.
• Also create a Cloud formation/ Terraform template to migrate the code pipeline for Development environment to Test and Production environment.
• As part of deployment AWS EC2 (Amazon Elastic Compute Cloud) instance, EKS (Amazon Elastic Container Service for Kubernets) instance will be created.
• Involve in creating Amazon Lambda functions to run quick serverless functionalities.
• Build the CloudWatch event to monitor the production pipeline performance and various matrix.
• Write the SQL queries, Stored Procedure and Function in Snowflake to pull the data based on users requirements.
• Involve in Data Migration from OnPrem systems which includes SQL Database, Teradata, Excel files to Cloud service like Snowflake and Amazon S3.
• Build the Attunity Tasks and Snowflake streams to migrate the data to cloud.
• Also involve in creating Tableau dashboards to visualize the data for the Quick Win presentation.
• Knowledge of Agile / Scrum methodology as a development process.
• Involve in Sprint Planning, Review and Retro, Daily standup and Weekly meetings to keep the Ongoing and Upcoming task on track.

Dec 2021- Present
Nov 2019 - Dec 2021

Business Systems Analyst

Netrush, Vancouver, WA

Key Responsibilities:
As a Business Systems Analyst, I am responsible for bridging the business needs of reporting high volumes of business data and insights into scalable, impactful tools (including the Netrush Brand Portal). Also work with the advisory and research teams to design data methods for capturing desired data, analyzing the captured data and creating methods for reporting findings to stakeholders. As part of process I Interact with users, stakeholders, Brand Managers, Marketing Analysts to collect User Requirements and translates into User Stores, Wireframes and Acceptance Criteria for development of the company wide Tools/Features.
As a Business Intelligence Developer, I am responsible for both ad hoc and organization wide data projects. Involving in writing SQL Quires and Procedures to extracts reports from Netrush database and Developing Power BI reports and dashboards to transfer insights in data to visualizations. I build Data Model using star schema for Analytics and Reporting purpose. As a day to day responsivities, I involve in reviewing reports and dashboard for fine tuning performance using Query Optimization techniques, applying reporting best practices. I develop ETL pipelines and activities in Azure Data Factory, to extract data from Netrush database and blob storage and apply transformation to convert production data into dimensional data and load into the reporting database. Also, I schedule these pipelines using Azure triggers to load SCD data and on daily intervals. In Azure Automation, I write Runbook scripts to Auto scale PowerBI premium tiers and schedule the Power BI reports refresh to keep upto date data at reporting server, also implementing incremental refresh of reports, trigger dashboards via email to stakeholders.

Primary Duties:
• Develop Azure ETL Pipeline/ Activities to extract data from production database/ Azure storage, transform into dimensional data and load into reporting database.
• Schedule these pipeline using Azure Automation.
• Writing SQL Quires and Procedures to extracts reports from SQL server database.
• Developing Power BI reports and dashboards to transfer insights in data to visuals.
• Implement calculation/formula to improve reporting performance.
• Dimensional Data Modeling using star schema.
• Review reports and dashboards for fine tuning performance using Query Optimization techniques, applying reporting best practices.
• Scheduling report refresh at keep current data and Implementing Incremental refresh to reduce refresh time.
• Trigger reports to different end users via email.
• Writing Runbook scripts in Azure Portal to Auto scaling Power BI preimium Tier on peek hrs.
• Interacting with users, stakeholders, Brand Managers, Marketing Analysts to collect User Requirements.
• Writing Features/Tool/Product Backlog Items with Description, User Stories, Acceptance Criteria.
• Involving in Standup, Sprint planning, User Review, Release and retro meetings.
• Creating Mockups/Wireframes as per user requirements.
• Testing the Tools/Reports/Dashboards and Creating Bugs if required.
• Creating Report, Tool release documents as a guidance for End Users and Stakeholders.
• Familiar with Azure DevOps/ VSTS for Agile process.

Oct 2018 - Nov 2019

Graduate Assistant

Athletics Department, Texas Tech University, Lubbock,TX

In this position, I developed and maintained the Texas Tech Athletic Department website. Created the tableau dashboards and stories to get insights and trends of expenditures and athletes’ performance in each game and also in academics. We created different kinds of charts, reports and motion charts using tableau. Deployed these dashboards to the Tableau public and used tableau bridge to keep the data UpToDate at the server side. To perform all these operations strong knowledge of database was required for writing SQL queries, PL/SQL procedures and functions to get data from various data sources. Experienced in logical and physical Data Modeling using normalizing Techniques, creating Tables, Views, Constraints, Index and Triggers using SQL and PL/SQL. Performed pivoting, clustering, created calculated and parameter fields to model the data according to the business needs. Tableau server was scheduled to send reports to via email on different time.

Also provided IT and desktop support for the athletics staff. As part of my duties, I created group policies, managed file shares and printers within the organizational unit. I also participated in the upkeep of the inventory by creating a tool to find missing equipment. Last, I participate in the deployment of software and patches through KACE and in the deployment of computer images through MDT via PXE boot.

Dec 2016 - May 2018

Software Engineer S2- Data Warehouse Analyst

EVRY India Pvt Ltd, Bangalore, India

As a Software Engineer, I was responsible for the Extraction, Transformation and Loading of data from multiple sources using Informatica ETL tools and SQL. To perform these duties, a deep understanding of the Data Warehousing concepts were required. Thus, I used Informatica Designer for designing and developing mappings from varied transformation logic using Expression, Source Qualifier, Filter, Router, Joiner, Lookup, Sorter, Aggregator and Update Strategy. I also designed ETL logic for Dimension Tables using SCD – Type1 and Type2. Last, I worked with heterogeneous data sources like DB2, Flat Files and Mainframe Files to accomplish the goals of the organization.

Aug 2014 - Aug 2016

Education

Master of Science in Computer Science

Texas Tech University,United States
August 2016 - May 2018

Bachelor of Engineering in Computer Science

Visvesvaraya Technological University,India
August 2011 - June 2014

Diploma in Computer Science

R N Shetty Polytechnic,India
July 2008 - June 2011

Skills

Programming Languages
  • PL/SQL
  • Shell scripting
Web Designing Skills
  • HTML5, HTML, CSS, JavaScript, PHP, XML, JSON
Database
  • Oracle 9i, SQL Server
Cloud Experience
  • Microsoft Azure : Data Factory, Azure Storage, Azure Automation,Runbooks
  • Amazon Web Services : S3, Lambda, SQS, SES, SNS, Step functions, Cloud Watch logs, Secret manager, Cloud Front, SSO,Cognito, EC2, EKS, ECR, Glue Etc..,
  • Other : Snowflake, Docker, Spark, Jenkins, Terraform, Gremlin, Git Actions
Business Intelligence Tools
  • Tableau, Tableau Desktop, Tableau Bridge, Tableau Server and Tableau Public, Power BI Pro, Power BI Desktop, Google Analytics, Looker, QuickBase, D3.js
Frameworks
  • Spring, Hibernate, Spring Boot (MVC)
Tools & Utilities
  • Attunity, Apache Spark, Informatica PowerCenter Client 9.5.1, Informatica Data Explorer (IDE), Informatica Cloud, SQL Developer and Toad, Power Exchange

Projects

Offline no shortage priority Analysis (Peterbilt)

Management spends time using EWI data to lookup defects and shortages on trucks to create priority lists to guide the offline flow to park trucks, the process of creating these lists is time consuming and the lists are not up to date by the time they reach the operators.
* I built a dashboard that will help the management to quickly see the number of defects on the trucks.
* Group and drill down to the trucks by part shortage category, line location , defect owner
* Identify and Prioritize low defect trucks and work on those so that those are ready to deliver
* See the truck parking location in the lot so that drivers can bring it to the plant for the rework.
* As this dashboard is real time which eliminates the manual work which is done before every shift.

Materials Parts Summary Analysis (Peterbilt)

This analysis was for the materials team to identify the number of trucks with part shortages by build day, and allows the plant to adjust build rates to reduce the impact of part shortages.
Key KPI were:
* Ability to see the Net on hand parts quantity which includes on hand and shipped quantity.
* Ability to see part utilization for the next 30 business days.
* Visibility to identity the part shortage using utilization calculated using build rate, required parts quantity based on scheduled build date/ chassis line order number
* Ability to see the shortage across all the manufacturing plants
* Ability to see chassis that are affected by part shortage, drill through part family

Truck Traffic Analysis (Peterbilt)

Project Overview & Objectives Dealer Development needs to implement the connected truck data into various analyses. Using the connected truck data, Dealer Development can accurately understand the level of truck activity in a given dealership location’s territory. This insight produces more precise recommendations for the need of more (or possibly less) of the following in a given dealership’s area/market:
* Number of dealerships
* Number of Service Bays
* Hours of Parts & Service operation
* Number of Mobile Service units
* Number of Outside Parts Salespeople

In this Project I worked as a Data engineer and performed below duties:

  • I conducted the POC to prove that processing millions of remote diagnostics data in python and finding the respected dealers using the user provided json territory files.
  • I developed the architecture to run the dealercd assignments on aws environment and using snowflake remote diagnostics and dealer territory files and store the results back to snowflake. And to implement business requirements.
  • I also did the cost estimation of the project which mainly includes AWS Services cost and Snowflake compute and storage cost.
  • Developed terraform templates for deploying the aws infrastructure on dev, test and prod accounts. Which involves AWS EC2, S3 bucket, Lambda function, SQS, SNS email
  • Performance of the process was a key part of the project so used python multiprocessing based on EC2 instance cores and daemon thread to sync logs to s3 while executing the script.
  • Developed Jenkins pipeline for deploying the infrastructure. And pipeline to build docker containers using the code from VSTS repo and deploy an EC2 instance and also pipeline to execute docker containers when file is uploaded to s3 and also on cron schedule.

Engine Reliability Analysis (Peterbilt)

Project Overview & Objectives The Engine Reliability Analysis (ERA) project was initiated as a directive from PACCAR's executive team. The MX13 EPA17 engine family is experiencing multiple catastrophic failure modes which have resulted in a significant increase in warranty costs. These issues also have a profound impact on the customer downtime, and the customer's perception of PACCAR's quality.
The ERA project is assisting with two primary use cases:
1.) Relative Risk of Failure - For each failure mode and each affected chassis, provide a projected probability of failure over the period of interest (e.g. next 12 months). Ranking by this probability provides a relative risk that allows the business to prioritize resources (dealership capacity, parts, warranty dollars, etc)

2.) Warranty Forecast - Projecting the number of failures that will occur within the warranty period. These projections are used to set the accrual amount for PACCAR's financial reporting. In addition to standard reliability projections, simulations are used to estimate the impact of the planned field interventions (campaigns, customer education, etc).

3.) Root Cause Analysis - Provide engineering insights with variables that are highly correlated with reliability to assist with generating root cause hypothesis. It is important to note that the analytics findings provide correlated variables, but these are not necessarily causal factors.

In this Project I worked as a Data engineer and performed below duties:

  • Designed the architecture running ML model on AWS using services like S3, SQS, SNS, EC2, ECR, Lambda and data from snowflake and parquet files
  • Designed snowflake er diagram for storing ML data, artifacts, build logs, database views.
  • Created the snowflake views using multiple tables/views from multiple databases for pulling the chassis, ops, EASOP and cab database and summarizing the remote diagnostics records for latest miles, engine revolution and engine hours.
  • Developed the terraform templates for deploying the aws infrastructure on dev, test and prod accounts.
  • Developed multiple Jenkins pipeline for deploying infrastructure, pipeline for deploying ML code and pipeline for running the models when file placed on S3 bucket.

Warranty Repairs Per one Hundred (R100) (Peterbilt)

Responsibilities:

  • I wrote a complex snowflake queries and views to calculate R100 for NIS/3MIS, Vehicle and Engine meeting at ATA and Parts level using Warranty, OPS and CAB database for the calculation R100 = total failures / total parts used.
  • Python scripts are written to read data from snowflake db and calculate R100 and store the results into project specific schemas.
  • Designed an architecture for running a script on schedule using AWS services (EC2, SNS, CloudWatch Event, S3, ECR)
  • Developed terraform template for deploying the AWS Infrastructure.
  • Developed Jenkins pipeline for running the deployments and running the scripts on schedule using cron job.
  • Used VSTS as a code repository and Jira for project management.
  • Developed a tableau dashboard for visualizing the R100 at ATA and Parts level, used various features of tableaus like charts, filters, actions, show hide menu, cross charts filtering, drill through.

Failure Mode Identification Project (Peterbilt)

I handled the data engineering apart of this project
Responsibilities:

  • Setting up the snowflake schemas, tables and views
  • Developed architecture for deploying the pipelines
  • Wrote python data util scripts for data scientists for reading and writing data to snowflake.
  • Developed flask application:
    * for downloading the claims and tagged failure mode for warranty meeting
    * for uploading validated files to snowflake by performing validation check
    * uploading and downloading Top ATA/Parts status files, Golden Set, Training data.
  • Deployed flask application on AWS EC2 instance using docker container.
  • Sync flask application log files to AWS S3 using APScheduler package.
  • Developed terraform template for deploying AWS infrastructure
  • Used VSTS as code repository and Jira for project management

Netrush Platform

This is a Netrush in-house retail domain project, focused on developing e-commerce platform. I am responsible for gathering user requirement and translating into development features. Also involve in testing tools.

Netrush Portal

This is a Netrush Analytical and reporting project, focused on developing reports and dashboards to get insights of sales, inventory, compliance, violations, pricing etc.,.

National Eligibility Database (NEDB Carrier Loads)

This is Insurance domain project for the client HMS (Healthcare Management System). We receive files on timely bases from Commercial Insurance carriers, Third party, Pharmacy Benefit Managers, Government Payers for which we do ETL (Extract Load and Transformation) and perform Data match using proprietary logic in order to Cost Avoidance and Recovery.

I was responsible for the Extraction, Transformation and Loading of data from multiple sources using Informatica ETL tools and SQL. To perform these duties, a deep understanding of the Data Warehousing concepts were required. Thus, I used Informatica Designer for designing and developing mappings from varied transformation logic using Expression, Source Qualifier, Filter, Router, Joiner, Lookup, Sorter, Aggregator and Update Strategy. I also designed ETL logic for Dimension Tables using SCD – Type1 and Type2. Last, I worked with heterogeneous data sources like DB2, Flat Files and Mainframe Files to accomplish the goals of the organization.

Team Activities:
  • Preparing Minutes Of Meeting
  • Giving internal training on Informatica PowerCenter.
  • Preparing Informatica training materials.
  • Updating the Track.


Academic Projects

Geospatial Visualization of Twitter Data

Introduction

This is an National Science Foundation funded research, it is focused on building a model for the researcher to collect tweets using Twitter API and analyze the trends in Twitter across Texas counties using hashtags to find the topics people are discussing every day. As hashtags are metadata, they help to recognize topic discussion at a first glance. Analyzing the evolution of topics over time shows how a discussion increases or fades away when new topics emerge. Discussions can also be suddenly switched to new topics. Human discussion behaviors can be studied through this type of analysis of topic evolution. Part of this analysis is frequency distribution of the hashtags which is calculated to highlight the use of the hashtags for each day. Another part of this analysis is applying topic modeling algorithms to tweets to bring out the generalized view point of the tweets. Over all tweets, conclusions are difficult to derive as many tweets consist of diverse topics; thus, topic analysis methods applied on a smaller set of tweets pertaining to hashtags shows more uniformity of thought in the tweets. These topics and hashtags are mapped to their respective county on a daily and weekly basis to find the co-occurring and identical discussion across the state of Texas. Further, to know the opinion of tweeters on these topics, tweets related to the hashtags are extracted and the sentiment, positive or negative, of each tweet is calculated using supervised machine learning algorithms. These positive and negative sentiment tweets are mapped to their respective county to find tweeter perspectives on the topics at each county level; thus, enabling researchers to factor in tweet analysis to their geospatial research of disease indicators, governmental policy issues, and social norms.

Application was built in Python using different packages like for UI {wxPython}, for charts and reports {bokeh}, for Analysis {pandas, SciKit-Learn, NumPy, SciPy, Plotly, NLTK, Gensim, Tweepy,json,urllib etc.}.



Movie Scope

Abstract

We implement a system through which we can analyze and predict the box office collection of the movies. We build the system in such a way that people from various backgrounds can also interact with the system and the data we have. Through the interaction they can also find hidden trends and insights which some of us miss. The data contains various variables which are compiled from various sources. The data can be extremely exhaustive but we have tried to collect the information essential for the analysis of trends and insights.

Introduction

Each year there are hundreds of movies which are released and being produced. The revenue of the movies depends on various factors like genre of the movie, critics review, quality of trailer, stardom of lead actor, studio of the movie being produced, director of the movie and various people included in movie making process. It can also depend on the current state of the economy, spending trends and habit of people, stock market and various other economic factors. The movie industry was worth more than 100 billion dollars in 2016[1]. And it is only expected to grow with number of new movies and stories and also the growing population. So it becomes useful to predict how much amount of money will one movie generate based on historic figures. In this paper we take into amount all the factors described above like genre, critics review, lead actor, studio etc. This demo helps to analyze and visualize the trends of the box office collection. End user can visualize the collections over the years based one the MPAA (Motion Picture Association of America) Rating's. Classify based on years, title name. Word cloud of top two movies help the user to know more about the movie summery, cast, genres. Review sentiment helps to know people opinion about the movie.

Click Here for Dashboard
Click Here for Github

A High Performance, Scalable Event-Bus System for Distributed Data Synchronization (Java Application)

Introduction

In this project, we basically focus to overcome the limitations of the existing techniques and our primary goal is to provide the support for fully transactional communication between the clients and the eventbus thereby making the communication more effective. Websockets replaces the native socket and multithreading concepts of java, which is the efficient way of establishing the pub-sub connection. Serialized the message delivery between publisher-eventbus-subscriber by using ‘Avro’ which is one of the serialization techniques used in hadoop.

Click Here for Github

Gnutella style P2P file sharing system (Java Application)

Introduction

This project is about implementation of Gnutella-style peer-to-peer (P2P) file sharing system, where signal peer act as both client and server with performing the file search within the specified network without using the Indexing server. I have implemented using java socket programming and multithreading concepts. When peer is started its respected server thread is started which runs in the background, given config file is read to find its peer and starts the client thread for each neighbor and when the neighboring peer gets online connection is established this connection is used to search file among the network. Server thread is used to search file within its local directory and store the result into the list once the local search is performed it reads the config file to start the client thread to all its neighbors. time-to-live (TTL) is attached to the message along with the filename need to be searched. TTL is used for preventing query messages from being forwarded infinitely many times. Client thread is used to send the file to search to servers which are connected to it. Server Download thread performs the downloading of file from the peer which is selected by the user. Same server and client threads are used for broadcasting the file modification message to all the neighboring peers which has that modified file.

Click Here for Github

Napster style peer-to-peer file sharing system(Java Application)

Introduction

I have used Socket Implementation of Napster-style peer-to-peer (P2P) file sharing system. The main idea of this project is to Create the indexing server which stores all the file details and bringing it online so that all the peers connect to it and register its files on the server. Once files are registered it provided with 2 option to search for the file in other peer or can exit the execution, if search is selected file name is asked and port number and sent to the server, sever will return the peer id’s which as the respected file and then client has to select the peer id from which file to be downloaded. Downloaded file will be stored into the requested peer’s shared directory. Every Peer can provide its shared director and can register its files so that other peer’s can search and download them. Every Client Server connection is done using Sockets, given the port number and the Server name client establishes a connection with the Server registering files and searching files among peers.

Click Here for Github

Visualizing time series data

Introduction

Representing the unemployment information for different states in USA based on the data set available Bureau of Labor and Statistics . This website allows the user to select the date range to generate a time series visualization.

Functionalities

  • User can zoom into time interval to get more insight.
  • User can visualize time series based on States.
  • User can select multiple states from the map for visualization.
  • User can compare unemplyment rate between all the states.
  • User can unselecte the selected states from the map.
  • User can reset the visualization.

Click Here for Github

Virtual PC Viewer (Android Application)

Introduction

Virtual PC Viewer is a remote control system which allows the user to view and interact with one computer (“remote desktop machine”) to cellular phones (“viewer”) using wireless internet connection. A viewer is provided on the cellular phone that enables the user to see and manipulate the desktop of various remote systems such as Windows operating system. The system to be accessed must be running a Virtual PC Viewer server and it must be attached to a network.


Description

Virtual PC Viewer is a client server mobile application and is an ultra-thin client system based on a simple display protocol that is platform independent. This software allows to communicate with the remote system and can have control on the peripheral devices like key board, mouse from the cellular phones. The remote system’s administration tasks can also be performed from the cellular phone. Popular uses for this technology include remote technical support and accessing files on one's work computer to the mobile device. Here the mobile device will act as client and remote machine is called server The user of Virtual PC Viewer has to register at client machine before connecting to server. Registered user can connect to the remote desktop machine through IP address and PORT number of remote desktop machine. After successful connection user can access remote desktop machine. Android native code is used to make client-server communication and better User Interface is provided to make the application user friendly.

Application

One of the primary application of this project is being able to access, operate a computer remotely from the mobile as if remote machine is in your hand, User can access virtual files from remote machine to mobile device.



Doodle Blink (Android Mobile Game)

Introduction

Now days the ease for playing software games has increased tremendously. Due to busy schedule and nuclear families people op to play software games rather than playing manually. Especially those which are quickly accessible via computers or mobiles. This game is one among such games which is fun filled, probabilistic in nature. The Doodle Blink is a skating game, the doodle will always be blinking, and instead the user needs to jump Doodle to skip the obstacles. When game gets over will get score in score board and it display only highest score.


Description

The Project “Doodle Blink” is a test of accuracy, strength and concentration. This is developed platform HTML5 and with the support of Phonegap. Doodle Blink game is the task completion game in which the doodle (person) will be always blinking, the doodle will be on the skating board will be with the doodle till the end of the task competition, for each of the task completion there will be the restricted time, the task contains the obstacles etc. Doodle blink a contiguous game in which doodle is always blinking and Instead the user needs to jump doodle to skip the obstacles and in the first level obstacles are less and speed is also less and we have to score 1000 points then will go for second level in this obstacles are more and speed is also more and for going the third level we have score 3000 points and in this level we have collect the positive points. In score board consists of highest top 4 score will be displayed.



Know before you buy (Web Application)

It is a website suggests user the products (Laptop or its accessories) from different brands based on his specifications (cost, brand, model). This website gives the complete idea regarding the latest products in the market.



File Converter (Java Application)

PDF is global standard for capturing and reviewing rich information. It is more secure and dynamic than ever. Basically PDF files are read only , which are not editable so we need to convert this file to some other format which provides more editing options one of the format is Microsoft word.

File Converter is a Windows system program. It supports 98/2000/XP/2007/2008 platform this software is developed on Java platform using Itext jar, This allows user to convert PDF file to document format file and vice versa and also PDF to TXT format. This software has the ability to retain the original layouts of the source file in the conversion.

Application:

  • Image extraction from PDF file.
  • Concatenation of two PDF files.
  • There is no problem in image processing.
  • Splitting of one PDF file to two PDF files.
  • Converting PDF file to Text file.


QuesZen

Introduction

In Education Centers the most required task is preparing question papers for internal, mid terms and final exams. Which need to meet in time and most secretive. Even university papers which are printed and sent to respective colleges have a chance of leak at last movement.


Description

QuesZen is developed on platform VB.net for frontend and MS Access as backend application and crystal report for generating the question papers into document format. The Software that generates the question paper by picking the questions dynamically from Question bank Database at the specified level of difficulty papers will be more secretive and it will be in time. It provides 3 levels of paper generation i.e. easy, medium and difficult levels. It allows selecting subject, level number of sections and number of questions of specified marks, it will generate different question set of each press of generate button by picking random question from the database and also allows you to change particular question from the generated paper, and provide you option save the 3 sets paper for each exam type.



Honors, Awards & Achievements

  • Geospatial Visualization of Twitter Data - Poster presentation at national workshop for NSF grant support.
  • Awarded competitive scholarship: 75% fee waiver by Texas Tech University.
  • WebMaster at Texas Tech India student Association.