Beruflich Dokumente
Kultur Dokumente
1 | April 2013
` 50/-
Technical Trend Socio-Business Intelligence using Big Data 11 Cover Story Big Data Systems: Past, Present & (possibly) Future 7
www.csi-india.org
CIO Perspective Deriving Operational Benets from High Velocity Data 28 Practitioner Workbench Programming.Learn ("R") 27 CSI Communications | April 2013 | 1
Security Corner
Information Security>>
ISACA and CSI Sign Memorandum of Understanding in Mumbai to Advance the IT Profession
and assurance in India, and to advance the IT profession, said Avinash Kadam, advisor to ISACAs India Task Force. The CSI President expressed the condence that this collaboration will grow beyond the cooperation between the CSI/ ISACA members, and will lead towards strengthening academia business and industry interaction. The CSI members immediate gain will be to benet from the ISACA, continuing professional development programme and access to the publications and learning resources. As a result of the MOU, ISACA will waive its new member fee for CSI members, who wish to join ISACA. CSI members will also receive a discount on ISACAs CISA, CISM, CGEIT, and CRISC certication exams. For more information on ISACA, visit www.isaca.org. To learn more about CSI, visit www.csi-india.org.
Mumbai, India (1 April 2013)Two major IT organizations in India have signed a memorandum of understanding to benet IT professionals throughout the country. The Computer Society of India (CSI) and ISACA yesterday signed an agreement that allows for mutual collaboration and knowledge sharing for the benet of the profession. ISACA is a global associations of 100,000 IT professionals who help enterprises ensure trust in, and value from, their information and systems. CSI has more than 100,000 members and 70 chapters in India. The MoU, signed by ISACA Director, John Ho Chi and CSI President Prof. S V Raghavan, notes that the organizations will advance the global IT profession in India, and the professional standing of ISACA and CSI members by: Strengthening the relationship among ISACA and CSI chapters in India Increasing awareness, use and adoption of the COBIT framework by CSI members Providing standard-setters, regulators and legislators with access to best practices, credentials and educational opportunities offered by CSI and ISACA Conducting joint educational events and research projects related to information systems governance, security, audit, and assurance issues in India ISACA is pleased to collaborate with CSI on this important mission: to promote information systems governance, security,
About CSI
Established in 1965, the CSI is Indias rst and the largest non-prot organization in the areas of information processing, computers, and communications. The mission of the CSI is to facilitate research, knowledge sharing, learning, and career enhancement for all categories of IT professionals, while simultaneously inspiring and nurturing new entrants into the industry and helping them to integrate into the IT community. The CSI is also working closely with other industry associations, government bodies, and academia to ensure that the benets of IT advancement ultimately percolate down to every single citizen of India. The CSI currently represents over 1, 00,000 members affiliated to 73 CSI professional chapters and about 562 CSI member institutions (including 499 CSI student branches), in different states and regions of India.
and education on information systems (IS) assurance and security, enterprise governance and management of IT, and IT-related risk and compliance. Founded in 1969, the nonprot, independent ISACA hosts international conferences, publishes the ISACA Journal, and develops international IS auditing and control standards, which help its constituents ensure trust in, and value from, information systems. It also advances and attests IT skills and knowledge through the globally respected Certied Information Systems Auditor (CISA), Certied Information Security Manager (CISM), Certied in the Governance of Enterprise IT (CGEIT), and Certied in Risk and Information Systems Control (CRISC) designations. ISACA continually updates and expands the practical guidance and product family based on the COBIT framework. COBIT helps IT professionals and enterprise leaders fulll their IT governance and management responsibilities, particularly in the areas of assurance, security, risk and control, and deliver value to the business. Participate in the ISACA Knowledge Center: www.isaca.org/knowledge-center Follow ISACA on Twitter: https://twitter. com/ISACANews Join ISACA on LinkedIn: ISACA (Official), http://linkd.in/ISACAOfficial Like ISACA on Facebook: www.facebook. com/ISACAHQ
Contact:
Faizan Aboli, Ketchum Sampark, +91 - 22 4042 5518, faizan.aboli@ketchumsampark. com Kristen Kessinger, ISACA, +1.847.660.5512, news@isaca.org
About ISACA
With more than 100,000 constituents in 180 countries, ISACA (www.isaca.org) is a leading global provider of knowledge, certications, community, advocacy,
Editorial Board
Chief Editor
Dr. R M Sonar
7 9 11 17 19 20 24
PLUS
Biji C L
Cover Story
Big Data Systems: Past, Present & (possibly) Future
Dr. Milind Bhandarkar
26 27 28 33 34
Practitioner Workbench
Editors
Dr. Debasish Jana Dr. Achuthsankar Nair
Resident Editor
Mrs. Jayshree Dhere
Technical Trends
Socio-Business Intelligence Using Big Data
Gautam Shroff, Lipika Dey & Puneet Agarwal
CIO Perspective
Deriving Operational Insights from High Velocity Data
Bipin Patwardhan and Sanghamitra Mitra
Research Front
Big Data Enabled Digital Oil Field
Pramod Taneja and Prashant Wate
Published by
Executive Secretary Mr. Suchit Gogwekar For Computer Society of India
Articles
Big Data
A Kavitha, S Suseela and G Kapilya
Security Corner
Avinash Kadam
IT Act 2000 Prof. I T Law Demysties Technology Law Issues Issue No. 13
Mr. Subramaniam Vutha
IT.Yesterday()
Please note: CSI Communications is published by Computer Society of India, a non-prot organization. Views and opinions expressed in the CSI Communications are those of individual authors, contributors and advertisers and they may differ from policies and official statements of CSI. These should not be construed as legal or professional advice. The CSI, the publisher, the editors and the contributors are not responsible for any decisions taken by readers on the basis of these views and opinions. Although every care is being taken to ensure genuineness of the writings in this publication, CSI Communications does not attest to the originality of the respective authors content. 2012 CSI. All rights reserved. Instructors are permitted to photocopy isolated articles for non-commercial classroom use without fee. For any other copying, reprint or republication, permission must be obtained in writing from the Society. Copying for other than personal use or internal reference, or of articles or columns not owned by the Society without explicit permission of the Society or the copyright owner is strictly prohibited.
35 37 38 39
Brain Teaser
Ask an Expert
40 41 42 43 44 45 46
CSI News
Published by Suchit Gogwekar for Computer Society of India at Unit No. 3, 4th Floor, Samruddhi Venture Park, MIDC, Andheri (E), Mumbai-400 093. Tel. : 022-2926 1700 Fax : 022-2830 2133 Email : hq@csi-india.org Printed at GP Offset Pvt. Ltd., Mumbai 400 059.
Regional Vice-Presidents
Region - I Mr. R K Vyas Delhi, Punjab, Haryana, Himachal Pradesh, Jammu & Kashmir, Uttar Pradesh, Uttaranchal and other areas in Northern India. rvp1@csi-india.org Region - V Mr. Raju L kanchibhotla Karnataka and Andhra Pradesh rvp5@csi-india.org Region - II Prof. Dipti Prasad Mukherjee Assam, Bihar, West Bengal, North Eastern States and other areas in East & North East India rvp2@csi-india.org Region - VI Mr. C G Sahasrabudhe Maharashtra and Goa rvp6@csi-india.org Region - III Prof. R P Soni Gujarat, Madhya Pradesh, Rajasthan and other areas in Western India rvp3@csi-india.org Region - VII Mr. S P Soman Tamil Nadu, Pondicherry, Andaman and Nicobar, Kerala, Lakshadweep rvp7@csi-india.org Region - IV Mr. Sanjeev Kumar Jharkhand, Chattisgarh, Orissa and other areas in Central & South Eastern India rvp4@csi-india.org Region - VIII Mr. Pramit Makoday International Members rvp8@csi-india.org
Division Chairpersons
Division-I : Hardware (2013-15) Prof. M N Hoda div1@csi-india.org Division-IV : Communications (2012-14) Mr. Sanjay Mohapatra div4@csi-india.org Division-II : Software (2012-14) Dr. T V Gopal div2@csi-india.org Division-V : Education and Research (2013-15) Dr. Anirban Basu div5@csi-india.org Division-III : Applications (2013-15) Dr. A K Nayak div3@csi-india.org
www.csi-india.org
Presidents Message
Prof. S V Raghavan From : president@csi-india.org Subject : Presidents Desk Date : 1st April, 2013
Dear Members
I deem it a great privilege to be at the helm of affairs of the Computer Society of India, and it is a great opportunity to be the President of the society, at a time when India is on the high growth path in electronics and computers. The recent policy declarations by Government of India National Telecom Policy, National Electronics Policy, and Electronics System Design and Manufacturing Policy open up tremendous possibilities for every Indian. From sensors to supercomputers, every area is open for innovation and rediscovery. Research and Development leading to Intellectual property generation and associated Human Resources Development for capacity building, in related areas are awaiting active participation from the CSI. Since 2010, India has integrated its knowledge generating institutions in the form of a National Knowledge Network (popularly known as NKN). In the same year Government of India, launched a project to take ber optic cable up to village panchayats through National Optical Fiber Network (popularly known as NOFN). Installation of broadbands everywhere, at speeds exceeding 10 Mbps / 100 Mbps / 1 Gbps, is making India the Best Connected Country. NKN has already connected close to 1000 national laboratories and institutes of higher learning in its fold, and moving towards the target of 1500 institutions. Virtual classrooms are slowly becoming the lifestyle in many of these institutions. NOFN spread and CSI spread across the country, seem to suggest tremendous opportunities to work together. Perhaps, the Division Chairs, SIG Chairs, Regional Vice Presidents, Chapter Chairs, and National Student Coordinator of CSI, would like to brainstorm and see what role CSI can play in this major change. CSI and Education had been synonymous, and perhaps can be a single focal theme for synergistic cooperation. The new Execom had its rst meeting on 31st March 2013. New Execom members and those who were continuing were excited about the days ahead. It is a real pleasure working with this team. I welcome all of them to this wonderful world of opportunity. As you all know, Shri H R Mohan and Shri Ranga Rajagopal have joined us as Vice-President and Treasurer respectively. Shri V L Mehta steps out as Treasurer after making sure that nances of CSI are stable and sound. Wonderful job indeed!
I would like to place on record the excellent work done by the outgoing team led by Shri Satish Babu. Many programs, MoUs, International relationships were handcrafted by them. Congratulations to your team Satish for giving a wonderful year to CSI. CSI signed a MoU with ISACA on 31st March 2013, for mutual cooperation. I am sure you will see the details elsewhere in this issue. Shri Satish Babu, will continue to represent CSI in SEARCC, BASIS, and ICANN. He has laid solid foundation between CSI and these entities, and will continue to strengthen it. I will support him in all his endeavors. I will represent CSI in IFIP General Assembly from now on. We have new web site and a new portal. Please use them and give feed back to CSI HQ. Web and Portal are the face of CSI, and hence our critical information infrastructure. Wonderful being with you, and I humbly seek your blessings and support. With best wishes, Prof. S V Raghavan President Computer Society of India
Editorial
Dear Fellow CSI Members, Welcome to CSI Communications Knowledge Digest for IT Community April 2013 issue. On behalf of CSI, as editorial panel members of CSI-C we are happy to convey to you that we have completed two years of editorship and are feeling privileged to bring about 1st issue of the third year. In this issue, we are covering articles on Big Data!, a buzzword everybody is talking about and trying to get hold of; a BIG overwhelmed response from our esteemed and fellow contributors proves that! We are still getting contributions; we could not accommodate all articles and would carry forward some contributions in next issue and request those who have sent contributions to bear with us. We are proud to tell you that we got a good number of contributions from our industry fellow professionals. This shows that Indian software companies are really serious about big data, are putting serious R&D efforts and must be looking at it as a big opportunity for India. We welcome and encourage our practitioners to contribute their valuable knowledge through CSI-C. A big thank to all our contributors. Hadoop - the name sounds familiar and immediately catches attention when somebody hears about big data. We start this issue with rst article under cover story section: Big Data Systems: Past, Present & (possibly) Future by Dr. Milind Bhandarkar, Chief Scientist at Greenplum, a Division of EMC2. He writes about big data, big data infrastructure, Apache Hadoop: its adoption and use cases and next frontiers for big data systems. Big Data A Big Game Changer, a second article in this section is by Shailesh Kumar Shivakumar, Technology Architect at Infosys Technologies, Bangalore. He writes about drivers and opportunities, impact and applications of big data, and mentions about how big data had a big impact and re-dened the way elections are fought in the recently concluded US elections. In technical trends section we have article by Dr. Gautam Shroff, Dr. Lipika Dey and Puneet Agarwal from TCS Innovation Lab titled: Socio-Business Intelligence Using Big Data. They describe how the fusion of social and business intelligence is dening the next-generation of business analytics applications using a new AI-driven information management architecture that is based on big-data technologies and new data sources available from social media. In research front section we have an article by Pramod Taneja and Prashant Wate of iGATE. They introduce readers to Oil and Gas domain, discuss need for digital oil eld enterprise platform, big data solution for digital oil eld with detailed functional overview in their article: Big Data Enabled Digital Oil Field.
CSI Communications | April 2013 | 6
We have three articles in article section. The rst article is on Big Data is by Kavitha, S Suseela and G Kapilya of Periyar Maniammai University. The second article is on In-Memory analytics by Prof Jyotiranjan Hota. He introduces in-memory analytics, application platforms, vendors, scope and benets, research challenges and future of in-memory analytics. Prof. Hota has been one of the regular CSI-C contributors. Opportunities always come with risks, however as a manager one should be knowledgeable about how to manage those risks. We have last article in this section on this topic by our regular contributor Avinash Kadam. The article covers in detail what a risk management professional is expected to be well versed with and describes these as ve key practice areas of risk and information systems controls. In Practitioner Work Bench section we have rst article under Programming.Tips () on Python: Programming Language for Everyone by Dr. Nibaran Das of Jadavpur University. The second article is by Prof. Umesh P and Silpa Bhaskaran of University of Kerala under Programming. Learn(R): R- StaR of Statisticians. In this section, they introduce a new language called R for the rst time. In CIO perspective, we have an article titled Deriving Operational Insights from High Velocity Data by Bipin Patwardhan and Sanghamitra Mitra of Research & Innovation, iGATE Mumbai, India, where they discuss about business drivers of big data and Data Stream Processing: genesis, introduction and implementation in various domains. In IT.Yesterday(), we have an article on founder of information theory and beloved father of information age Claude Elwood Shannon titled Birthday Tribute to the Most Inuential Mind of 20th Century by Research Scholar Biji C L from University of Kerala. There are other regular features such as Security Corenr, Brain Teaser, Ask an Expert and ICT News Brief in March 2013 in Happening@ICT, CSI reports, chapter and student branch news and various calls. Remember we look forward to receiving your feedback, contributions and replies as usual at csic@csi-india.org. With warm regards, Rajendra M Sonar, Achuthsankar S Nair, Debasish Jana and Jayshree Dhere Editors
www.csi-india.org
Cover Story
Apache Hadoop
The GFS and MapReduce papers motivated Doug Cutting, creator of an open-source search engine, Apache Lucene, to rearchitect the content system of Lucene, called Nutch, to incorporate a distributed le system, and MapReduce programming framework for tasks of crawling, storing, ranking, and indexing web pages so that they could be served as search results by Lucene. These developments were noticed by engineers and executives at Yahoo, which was then struggling to scale its search backend infrastructure. Yahoo adopted Apache Hadoop in January 2006, and made signicant contributions to make it a scalable and stable platform. Today, Yahoo has the largest footprint of Apache Hadoop, running more than 45,000 servers managing more than 370 Petabytes of data with Hadoop[7]. Being an open source system, licensed under the liberal Apache Software License, governed by the Apache Software Foundation, meant that Hadoop could be freely downloaded and deployed in any organization, modied and used without any hard requirement of having to contribute the changes back to open source. The scalability and exibility of Apache Hadoop prompted growing Internet companies such as Facebook, Twitter, and LinkedIn to adopt it for their data infrastructure, and contribute
modications and usability enhancements back to the Apache Hadoop community. As a result, the Hadoop ecosystem grew rapidly over the years. Today, there are more than 20 components in the Hadoop ecosystem that are developed as open source projects under the Apache Software Foundation, and several hundred proprietary and other open source components. Some of the popular components in the Hadoop ecosystem, apart from Hadoop Distributed File System (HDFS), and MapReduce, include Hive, A SQL-like language that translates to MapReduce; Pig, an imperative data ow language that generates MapReduce jobs to execute the data ow; and HBase, a NoSQL Key-Value store that uses HDFS as its persistent layer. HBase is based on a paper describing another Google infrastructure component, Bigtable, which was published in 2006[8]. While Hadoop, today, has become the de facto platform for analyzing Big Data, challenges remain in making it accessible and improving its ease of use, thus making it a rst-class citizen of data infrastructure managed by IT professionals. The MapReduce programming paradigm is not particularly easy to use for data analysts, and commonly used business intelligence tools do not interoperate with interfaces provided by Hadoop today. To overcome these challenges, a number of data warehousing system vendors, such as Teradata, Oracle, IBM, EMC2/Greenplum, and others offer connectivity with Hadoop platforms. There are even efforts towards unifying SQL-based OLAP platforms, such as Greenplum, with Hadoop[9]. A number of Hadoop distributions have emerged over the years, improving manageability of Hadoop infrastructure. These include Cloudera, Hortonworks, MapR, EMC2/ Greenplum, IBM BigInsights, Microsoft HDInsights, etc. In addition, there is an increasing number of Big Data Appliances; hardware platforms that are integrated with Hadoop distributions, including Oracle, Teradata, and EMC2/Greenplum.
others have been almost fully digitized. Until recently, the data these organizations used to collect were stored in archival systems, mostly for regulatory compliance purposes. However, there is a growing realization across these organizations that this data can be utilized for gaining competitive advantage, increasing process efficiencies, and improve customer experience. In a recent study conducted by Tata Consultancy Services (TCS)[10], over 50% of organizations surveyed are using Big Data technologies, and many of them predicted more than 25% gains in returns on investment (ROI), mostly from increased revenue. The exibility of these Big Data systems to combine structured datasets (51%) with semi-structured datasets (49%) has been cited as enabling advanced analytics capabilities. In addition, while most of the organizations use data that is available internally (70%) within those organizations, availability of external data, such as from twitter and other social media, allows them to perform better customer behavior analysis. The 3Vs, volume, velocity, and variety of data, along with need to develop agile, data-driven applications, implies that the humans analyzing, detecting patterns, and making sense of data need to have a rich toolset at hand. Traditional data exploration, visualization, business intelligence, and reporting tools are being adapted to co-exist with these new Big Data technologies. Advances in machine learning algorithms and methods, as well as abundant processing power, have democratized deep and predictive analytics to be used in any average IT department. Open source languages for statistical analysis and modeling, such as the popular R language[11] and a newcomer such as Julia, as well as emerging machine learning frameworks, such as scikit-learn in Python[12], Apache Mahout for Hadoop[13], and In-Database deep analytics library, MADlib[14] have attracted attention of developers and users for developing machine-learning powered applications based on large and diverse datasets. These new platforms, languages and frameworks have challenged several predominant practices in the enterprises. Traditional data governance practices, including access control, provenance, retention, backup, mirroring, disaster
recovery, security, and privacy, are struggling to cope with organizations ability to store and process massive amounts of diverse data. Over the next few years, one should expect best practices for data governance, and associated technologies to emerge and become commonplace.
Continued on Page 16
CSI Communications | April 2013 | 8 www.csi-india.org
Cover Story
Big data also provides the following opportunities: Improve productivity and innovation McKinsey predicts an increase in job opportunities ranging from 140K to 190K Uncover hidden patterns and rapidly respond to changing scenarios. Multi-channel and multi-dimensional information aggregation Data convergence
The main value that can be derived from Big Data is by aggregating vast amount of data integrated from various sources. Following diagram shows various technologies used in Big Data:
Traditional search, sort, and processing algorithms would not scale to handle the data in this range, and that too most of it being unstructured. Most of the Big Data
Stock advises based on huge amount of stock data analysis, unstructured data like social media content etc. Credit worthiness analysis by analyzing huge amount of customer transaction data from various sources Pro-active fraudulent transaction analysis Regulation conformance Risk analytics Trading analytics Better analysis of supply chain data and touch points across Omnichannel operations Customer segmentation based on previous transactions and prole information Analysis of purchase patterns and tailor made product offerings Unstructured data analysis from social media, multi-media to understand the tastes, preferences, and customer patterns and do sentiment analysis Targeted marketing based on user segmentation Competitor analysis Mining of customer location data, call patterns. Integrate with social media to provide location based services like sale offers, friend alerts, points-ofinterest suggestions etc. Geo-location analysis : Effective drug prescription by analyzing all structured and unstructured medical history and records of the patient Avoid un-necessary prescriptions Risk analysis of customer Analyzing cross-sell and up-sell opportunities based on customer spending patterns
Retail industry :
and designing effective policies to persuade them The data analysis included mining customer data, proling them and sending targeted campaign mails to inuence their decision. The analysis also provided crucial insights about the voters who are most likely to switch sides and the required triggering points for the switch. The team built persuasion model with predictive analytics to nd out the probability of persuasion among population various geographies.
Analyzing the Big Data was the key differentiator in swinging a good percentage of voters and predicting the results with a greater condence.
Market Opportunity
Big Data offer bigger opportunities. Here is a snapshot of some of the predictions done by market research rms in this regard: IDC predicts the Big Market to grow to $16.9 Billion by 2019 Digital reasoning estimates that Big Data market would be worth $48.3 billion in 2019
Mobility:
References
[1] http://blogs.wsj.com/digits/2009/ 05/18/the-exploding-digital-universe/ Risk analytics [2] h t t p : //w w w . f o r b e s . c o m /s i t e s / tomgroenfeldt/2012/01/06/big-databig-money-says-it-is-a-paradigmbuster/ [3] http://www.emc.com/about/news/ press/2011/20110628-01.htm [4] EMC link [5] http://online.wsj.com/article/SB10001 4241278873233532045781266711241 51266.html [6] h t t p : //w w w . e w e e k . c o m /c /a / Application-Development/Big-DataMarket-to-Grow-to-169-Billion-by2015-IDC-118144/ [7] h t t p : //w w w . f o r b e s . c o m /s i t e s / n e t a p p/2 0 1 2 / 1 1 /0 6 / b i g - d a t a election-surprising-stats/ n
Health care:
Insurance:
Shailesh Shivakumar is a technology architect at Infosys with over 11 years of industry experience. His areas of expertise include Java Enterprise technologies, Performance engineering, Enterprise portal technologies, user interface components and performance optimization. He was involved in multiple large-scale and complex online transformation projects for marquee clients of Infosys. He also provided on-demand consultancy in performance engineering for highly critical projects across various units. He is a regular blogger at Infosys Thought Floor, and many of his technical white papers are published in Infosys external site and in Infosys Lab briengs journal. His blog also was listed in Most popular category recently. He also heads a centre-ofexcellence at Infosys. He also holds numerous professional certications including Sun certied Enterprise Architect (part 1), Sun certied Java programmer, Sun Certied Business component developer, IBM certied Solution Architect Cloud computing, IBM Certied Solution Developer IBM WebSphere Portal 6.1, and many others.
www.csi-india.org
Technical Trends
What is BigData?
The term BigData has become the latest buzzword in the IT industry, much as Cloud Computing began to elicit interest a few years ago. As in the case of the latter, we submit that BigData, is a metaphor for a few signicant technology, social business convergences: Popular interest in cloud computing was fuelled by the emergence and eventual conuence of web-based social applications, software as a service, infrastructure as a service, and nally platforms as a service. In a similar fashion, BigData is essentially the convergence of technology advances in articial intelligence emanating from search and online advertising, along with the development of new architectures for managing extremely large web-scale data volumes, exemplied by the now popular Hadoop stack. Along with the means to process vast quantities of unstructured data, we also nd that the data itself is now readily available: Vast volumes of consumer conversations on social media;, such as Twitter, are free for all to access, and the rest are rapidly becoming a valuable commodity available for purchase from Facebook, Linkedin, etc. In this article we describe a number of Socio-Business applications that exploit these new data sources, and are of potential interest to large enterprises. Moreover, we nd that each of these applications involve the fusion of information from social media with internal business data, the extraction of knowledge from web-sources, the application of articial intelligence techniques in some fashion, and/or the exploitation of BigData-inspired datamanagement architectures.
Supply-Chain Disruptions
The recent natural disasters that struck Japan in 2011, i.e., the earthquake, tsunami, and subsequent release of nuclear radiation, clearly had a devastating effect on the the Japanese population and economy. At the same time, the effects of these events were felt around the world; in particular, they led to major disruptions in the global supply chain for many industries, from semiconductors to automobiles and even baby products. The Japanese earthquake was a major event of global signicance, followed closely in the global media on a daily basis; hopefully a fairly rare black swan event. However, many adverse events of a far smaller signicance occur daily across the world. Such events are mainly of local interest only. Further, public interest in the event may last but a day or so, while its economic impact may last much longer: Take the example of a re in a factory. There are, on the average, in the range of ten or so major factory res in the world every day. Similarly there are labour strikes that disrupt production. Most of these events affect a very small
locality, and may not even reach the local news channel, and certainly not global ones. Further, any public interest, however localised, in the event may last a few hours or at most a day. Nevertheless, if the factory affected is a signicant supplier to a major manufacturer half-way around the world, this relatively minor event is possibly of great interest to the particular enterprise that consumes its product! It is observed that the manufacturers notice such news about their suppliers, when they encounter shortage in supply, which is usually a few days or sometime a week later. If however technology can help them notice this earlier, they will have more time to make alternate arrangements. Interestingly, it has been found that many of these events, even the ones with extremely local impact, nd their way fairly rapidly into social media, and in particular Twitter. Used for social-networking in over 200 countries with over 500 million tweets a day, Twitter turns to also be rich source of local news from around the world. Many events of local importance are rst reported on Twitter, including many that never reach news channels. Fig. 1 describes the overall architecture for listening to events from social media that we have used both for detecting adverse events as well as for listening to the voice of the customer as described in the next Section. In[5] we have proposed an architecture that enables a large enterprise to monitor potential disruptions in its global supply-chain by detecting adverse events, by monitoring Twitter streams. In[4] we have described how such events can be efficiently detected using machinelearning techniques, from amongst streams of unstructured short-text
messages (tweets) arriving at a rate of tens of messages per second. In contrast with the larger volumes that follow events of wider signicance, there are often only a few tweets reporting each such event; the few tweets that happen to report the same event, are then correlated. Next, as described in[5] and, the impact of the detected event to the enterprise in question can be assessed, by fusing the detected external event with internal data on suppliers.
services, or even specic pre-dened features of a product or service. The emphasis is on volumes, and on tracking the overall aggregate positivity / negativity associated with the set of concepts one is interested in. Source selection is broad and channel based; thus one might choose to focus on say Twitter, a Facebook page, and selected blogs, as well as analyze the variation across these. Since sentiment is noisy and varies rapidly, it is also aggregated temporally; thus the timescales of aggregate sentiment analysis are in the range of days and weeks. Social-media-based brand sentiment analysis is cheaper and faster than traditional survey-based techniques such as Nielsen market-surveys; it also reveals results sooner. Thus, sudden and signicant changes in sentiment about a brand can be detected faster, such as that which took place when tropicana changed its packaging a few years ago, which was followed by strong negative consumer sentiment. However, the jury remains out as to how often these aggregate sentiment gures bring novel insights as compared to traditional measures. The fact is they need to be time-averaged to make any sense; thus ner-grained approaches are needed to enable more real-time response, and detect emerging problems that by themselves may not change the aggregate sentiment signicantly, at least at rst, and that too only if not addressed in time. Listening to consumer sentiment on social platforms has recently become almost a commodity offered by a number of commercial services, such as Radian61 and others. Opinion-mining techniques for extracting sentiment from text are used in such tools. The initial insight that is most often sought through the adoption of a listening service is the ability to monitor brand perception, i.e., whether consumers at large are saying positive or negative things about ones brand, product, or service. 2. Complaint Analysis, in contrast with brand sentiment analysis that casts its net wide, complaint analysis tries to focus on actual customers. Thus, the sources for such analysis are either direct customer feedback through call-centers or email, or when it comes to social-media, the input is carefully ltered so as to ensure the presence of indicators such as I bought, my car, etc., www.csi-india.org
making it highly likely that the writer is in fact a customer, either of ones own product or that of a competitor. Next, such complaint analysis aims to analyze the text written by customers to detect which aspects of a product or service they are having difficulty with. This requires a deeper level of natural language processing than, say, aggregate sentiment analysis: Consider the statement Ive been having trouble with my new [carbrand], not only did the transmission give way in the rst month but there was a signicant delay in getting it changed. Clearly it is a negative statement about the car brand, and even its transmission, which basic sentiment analysis can easily discover. However, deeper text processing can further discern what exactly is wrong with the transmission, and aggregate such difficulties across a large volume of customer feedback along various dimensions. As a result, if the concept of say, transmission giving way including its linguistic equivalents, is showing up in signicant numbers, then this becomes an issue to ag to product engineering. On the other hand, the fact that the supply of spares of various types are delayed, including transmission parts etc., gets aggregated at a different level of say delayed parts, and is escalated to those responsible for after-sales services. The deeper degree of text processing required for complaint analysis requires
in communications regarding particular new terms, such as iPad, so as to detect emerging problems even if they are not part of a known categorisation or ontology. Summarising, causal feedback analysis restricts the source to customer feedback, analyses the content in depth, and aggregates results over a period of time, for example on a weekly or monthly basis. Most importantly, in contrast with brand sentiment analysis, complaint analysis often results in directly actionable intelligence that can be passed on to the concerned division in the enterprise. Fig. 2 describes our architecture for ontology driven opinion mining from unstructured customer feedback, which is described in more detail in[2]. 3. Early problem detection, again listens to consumers at large. However, unlike aggregate sentiment analysis, the aim here is to quickly detect new problems being faced by consumers. For example, a new website design might be awed, leading to consumer frustration; a new policy on a banking service may be leading to angst and outrage, or a major competitor might be luring customers away. Increasingly, consumer conversations that might point to such events are taking place in the open, on Twitter. However, as in the case of detecting potential supply-chain disruptions, the stream of tweets needs to be ltered to rst focus only on consumer complaints, and then processed to extract information on the actual problem being faced. However, the situation here is technically more challenging than say, factory events, since distilling consumer versus non-consumer events is less accurate than discerning factoryre or labour-strike events. Further, the Fig. 2: Ontology-driven opinion mining nature of the information that one seeks to extract ontology-driven causal analysis, which need not be known in advance. Thus, involves some level of parsing as well as while a domain ontology can help classify learning, and exploiting domain ontology. events to a certain extent, the sudden Additional techniques required include arrival of a comment on why isnt there an trend analysis, whereby sudden spikes iPad application to access my ... account
1
like there is for [competitor]; it may well be that iPad does not yet gure in the domain ontology. Still, this problem needs to be detected and classied in some manner so that appropriate action can be taken. New problem detection is as yet difficult to completely automate: Instead, as in the example above, it is better to bring a human in the loop when required; of course, automatically guring out when to do so is equally important as well.
Competitive Intelligence
Competitive intelligence is aimed at assessing risks and opportunities in a competitive environment, before they become obvious. It is used by organisations to compare themselves with their peers (competitive benchmarking), to identify risks and opportunities in their markets, and to pressure-test their plans against market response (war gaming), which enable them to make informed decisions. Competitive intelligence comprises the tasks of dening, gathering, and analysing intelligence about the industry in general, along with specic knowledge like products, pricing, marketing strategies, and much more about competitors. The information gathered allows organizations to realize their strengths and weaknesses. Acquisition and analysis of events falling under competitive intelligence category is a highly specialised activity. Competitive intelligence can be broadly classied into two categories depending on whether it is used for longterm planning or short-term planning. Strategic Intelligence (SI) focuses on long term issues that analyze a companys competitiveness over a specied period in future. The main focus of analysts here is to forecast, where the organization should be positioned few years hence, and to identify strategies to convert this into a reality. This analysis primarily involves identifying weaknesses, and early warning signals within the organization. Tactical Intelligence on the other hand focuses on providing information that can inuence short-term decisions. Most often, this is related to analysis of current market share and the competition landscape. This kind of intelligence directly affects the sales process of an organisation. Tactical intelligence can be further categorized as: (i) Brand related: provides information about popularity of competitors in terms of their products
Type of Cometitive Intelligence Event People events Competitor strategies. e.g. technology investment Consumer sentiments Promotional events and pricing Related real-world events
Web Source News, company web-sites News, Discussion Forum, Blogs, patent sites Review sites, social networking sites Twitter, Facebook News Twitter, Facebook
or brands as a whole, which products are moving in the market, market share of competitors. Consumer sentiments related to the organization and its competitors also belong to this category. (ii) Pricing related: provides knowledge about prices of competitor products. (iii) Promotions related: provides information about promotion strategies and kind of promotional activities that are adopted by competitors. (iv) Organizational: provides information about competitors like their work force structure, internal shift in focus or vision, success or failure of their trials, new product launches, technology investments etc. all contribute towards building a prole of competitors that can be useful to organizations. The table in Fig. 5, presents an overview of how different types of web-content can contribute towards compiling tactical competitive intelligence reports for an organization. A detailed treatment of how competitive intelligence can be extracted from social media is treated in[3]
to competitive intelligence. Discussions on different forums and blogs can provide crucial insights when analyzed in proper perspective. Using Google search trends for competing products and services can also be a good source of competitor intelligence. It is essential to dene a set of processes to gather information, converting it into competitive intelligence, and then channelize it for consumption in business decision making. Usability and actionability of the information gathered are two critical factors in determining its relevance. Information gathered from the Web is unstructured in nature, and therefore not immediately machineinterpretable. Handling inaccuracies, redundancies, and volumes are other challenges. Appropriate knowledge management techniques are required to ensure that analysts have access to all relevant information to without facing information overload. Fig. 4: Analyzing competitor promotions Given the large volumes from social media of information received in chunks of information in a structured a digitized format, natural language form that can be consumed even by processing, text mining, and statistical machines. Information and relation reasoning play signicant roles in mining techniques have been successfully automating the process of content applied to extract signicant entities, and assimilation. A host of specialized tools their roles and responsibilities in an event are also available to aid some of these along with event details like name, time, tasks. News analytics is a well-established location, and description of event. The research area dedicated to analysis and structured information extracted from the organization of news articles received from news articles can be further consumed by different sources, to predict the political, a reasoned to draw inferences. nancial or social impacts of these Social media content on the other events. Extracting specic events that hand can contribute very effectively can contribute to competitive intelligence towards gaining tactical intelligence. can be considered as a sub-task of news Tracking twitter and Facebook content analytics. Classication techniques are
employed to classify news articles into broad categories like political, economic, sports, market information, entertainment etc. Article summarization techniques are often used along with this to provide the key content of articles. Clustering news articles based on content is also an oftused technique to reduce information overload. Intelligent cluster visualizations help in easy assimilation of content. One of the key challenges here is to identify those events which can be assessed for their impact on past, present or future performance of an organization. All impacts are not measurable. For example, it is difficult to measure the impact that a new technology may have on the future market or the effect of a new chief appointed by a competitor or even the news about an important acquisition by a large company. News events typically comprise a major chunk of information used to gain strategic intelligence. Information and relation extraction techniques from text mining are also gaining popularity in news analytics, since they can further help in extracting specic
www.csi-india.org
generated by competitors, can provide fairly accurate data about promotions run by them. Twitter and Facebook also abound on consumer sentiments about products and services or a brand. Text classication techniques are widely used to classify social media messages, into pre-dened categories like status updates, sentiment, and opinion, consumer support systems, news, promotions and campaign, and others. Further categorization or labelling of content is also possible based on the named-entities present in these. Classication of social-media content into pre-dened categories like above helps in ltering the relevant from irrelevant. Traditional classication techniques using the Bag-of-Words do not perform very well on short messages like these. Rather a set of domain-specic features like authors prole, ReTweets, @userMentions etc., help in classifying the text to a predened set of generic classes such as News, Events, Opinions, Deals and Promotions, and Customer Support. A classied text can be further tagged or associated with product or service labels, brand names, action categories etc. using domain ontology. Natural Language Processing tools like Named Entity Recognition is also applied to identify dates, money-values, store names or locations etc. The assigned class and product labels along with the complete set of information extracted can be used to generate a promotion map, which can depict category-wise promotions for different products regionwise and time-wise. Fig. 4 depicts the process ow for the same.
of data from multiple sources, requires intelligent master data management techniques. Fusion systems need to judge feasibility and relevance of the merging different types of data. Visualization of the results to deliver the correct insights is yet another complex task. While much of this work is also human-driven today, analytical systems that can fuse competitive intelligence reports and structured data at the right granularity are being developed for different sectors. Machine learning techniques are major contributors to the design of fusion systems. These systems can be made to learn from human interactions with reports and data. The marketing division is one of the most prolic users of social media. Consequently, they can also maximally benet from competitor promotion information. Most companies have a predened static promotion calendar. This calendar is reviewed from time to time, usually on a quarterly basis. The review is most often entirely against the companys own performance, without information about competitor actions used in a structured way. Promotion event maps created from social media can be used by the marketing analysts to get a near real-time view of competitor activities, analyze the companys performance against the backdrop of these and thereby take corrective actions, if necessary. Joint analysis of sales data and competitor promotion events, can provide valuable insights about how competitor promotions affect sales. For example, a dip in sales data can be linked to reports about aggressive promotions by competitors, new product launch in the same category, pricerise announcements or sudden rise in negative brand sentiments. Similarly, rise in sales can be linked to rise in positive brand sentiment or price rise announced by competitor. Given that there may not be a single well-dened factor that can be marked as responsible for an event, automated systems can do a good job of correlating all that is relevant based on attributes like time of the year, product, brand or region. Pattern mining on large volumes along with human annotations, as input can be utilized to learn better correlations. Finally, machine learning driven competitive intelligence
systems can also be used to design predictive models that can predict future performances based on series of present and past events.
References
[1] Kemal A Delic and Umeshwar Dayal. The rise of the intelligent enterprise. Ubiquity, 2002 (December): 6, 2002. [2] Lipika Dey and Sk Mirajul Haque. Opinion mining from noisy text data. International journal on document analysis and recognition, 12 (3): 205 226, 2009. [3] Lipika Dey, Sk Mirajul Haque, Arpit Khurdiya, and Gautam Shroff. Acquiring competitive intelligence from social media. In Proceedings of the 2011 Joint Workshop on Multilingual OCR and Analytics for Noisy Unstructured Text Data, page 3. ACM, 2011. [4] Saurabh Sharma, Puneet Agarwal, Rajgopal Vaithiyanathan, and Gautam Shroff. Catching the long-tail: Extracting local news events from twitter. In International Conference on Weblogs and Social Media, June 2012. [5] Gautam Shroff, Puneet Agarwal, and Lipika Dey. Enterprise information fusion for real-time business intelligence. In Proceedings of the 14th International Conference, Fusion 11, 2011. n
Dr. Gautam Shroff is Vice President & Chief Scientist, Tata Consultancy Services and heads TCS Innovation Lab in Delhi, India. As a member of TCS Corporate Technology Council, he is involved with recommending directions to existing R&D efforts, spawning new R&D efforts, sponsoring external research, and proliferating the resulting technology and intellectual property across TCS businesses. Prior to joining TCS in 1998, Dr. Shroff had been on the faculty of the California Institute of Technology, Pasadena, USA and thereafter of the Department of Computer Science and Engineering at Indian Institute of Technology, Delhi, India. He has also held visiting positions at NASA Ames Research Center in Mountain View, CA, and at Argonne National Labs in Chicago. Dr. Shroff completed his B.Tech (Electrical Engineering) from the Indian Institute of Technology, Kanpur, India, in 1985 and Ph.D. (Computer Science) from RPI, NY, USA, in 1990. Dr. Shroff taught a course Web Intelligence and Big Data on Coursera as well as at IIT and IIIT and the URL is https://www.coursera.org/course/bigdata . Dr. Lipika Dey is a Senior Consultant and Principal Scientist at Tata Consultancy Services, India. She heads the Web Intelligence and Text Mining research group at Innovation Labs, Delhi. Lipika's research interests are in the areas of content analytics from social media, social network analytics, predictive modeling, sentiment analysis and opinion mining, and semantic search of enterprise content. Her focus is on seamless integration of social intelligence and business intelligence. She is keenly interested in developing analytical frameworks for integrated analysis of unstructured and structured data. Lipika has a Ph.D. in Computer Science and Engineering from IIT Kharagpur. Prior to joining the industry in 2007, she was a faculty member in the Department of Mathematics at Indian Institute of Technology, Delhi, from 1995 to 2006. She has several publications in International journals and refereed conference proceedings. She is a Program Committee member for various International Conferences. Puneet Agarwal is a Scientist at Tata Consultancy Services Ltd. He heads Data Analytics and Information Fusion research group at TCS Innovation Labs, Delhi. Puneets research interests include applied research in data-mining on time-series and graph data with a focus on distributed parallel processing. He has been working in TCS for about 15 years and before joining TCS Innovation Labs in 2004, he worked as a technical architect in various mission critical projects in Logistics and Shipping domain. He has published many a research papers in various international conferences on Information Fusion, Software Agility, Collaboration, and Model Driven Interpretation. Puneet has a B.E. Degree in Mechanical Engg from NIT Trichy.
References
[1] [2] IT Glossary, Gartner Inc, http://www. gartner.com/it-glossary/big-data/ The Digital Universe in 2020: Big Data, Bigger Digital Shadows, and Biggest Growth in the Far East, http://www.emc. com/leadership/digital-universe/iview/ index.htm, December 2012 Metcalfes Law Recurses Down the Long Tail of Social Networks, http://vcmike. wordpress.com/2006/08/18/metcalfesocial-networks/, April 2006 We knew the web was big, http:// googleblog.blogspot.com/2008/07/weknew-web-was-big.html, July 2008 The Google File System, http://research. google.com/archive/gfs.html, October
[3]
[4]
[5]
Dr. Milind Bhandarkar was the founding member of the team at Yahoo that took Apache Hadoop from 20-node prototype to datacenter-scale production system, and has been contributing and working with Hadoop since version 0.1. He started the Yahoo Grid solutions team focused on training, consulting, and supporting hundreds of new migrants to Hadoop. Parallel programming languages and paradigms has been his area of focus for over 20 years, and a topic of his PhD dissertation at University of Illinois at Urbana-Champaign. He worked at the Centre for Development of Advanced Computing (C-DAC), National Center for Supercomputing Applications (NCSA), Center for Simulation of Advanced Rockets, Siebel Systems, Pathscale Inc. (acquired by QLogic), Yahoo and LinkedIn. Currently, he is the Chief Scientist at Greenplum, a division of EMC2.
www.csi-india.org
Research Front
Introduction
Fig. 1 : Functional overview big data enabled digital oil eld Source: iGATE Research
extremely large data sets on commodity servers. Hadoop based solutions allow storing, processing, and analyzing these humongous logs on near real time basis. The crux of the solution involves processing raw data in its native format to create aggregated views along with understanding of its relationships and patterns and thereby derive meaningful insights for quick decision-making related to reservoir & optimizing the data exploitation using Map Reduce paradigm. There is a widely acceptable adoption of Hive, which is a scalable data warehouse solution available on Hadoop, with HiveQL as a query mechanism which is similar to SQL syntax. Hive internally generates map-reduce jobs that can be executed on Hadoop clusters. Hive in-turn allows overcoming the learning curve associated with the Map-Reduce code generation.
production phase, leveraging modeling/ simulation techniques, and ready-todeploy KPI congurations The high-level functional overview is stated below:1. Fetch the customers E&P from various well in different phases Each oil well generates around 10TB of data and in a reservoir there are multiple wells to be drilled and explored 2. This massive volume of multi structured logs is stored on a Hadoop infrastructure 3. Data Processing is most important step for data preparation in a manner which is less time consuming activity. Hadoop is an ideal solution which can be used for converting the unstructured data to structure format, perform cleansing and store in unied Hive structures. 4. The Digital Oil Field provides PPDM compliance models for ease of integration and portability post standardization into a Digitized Platform 5. The quality of data on the Digitized Platform is veried by the stakeholders 6. Complex analytics and event processing is performed to nd the drilling patterns, infer the lithology content based on various parameters 7.
of the Oil well logs. In turn provide adaptors to 3rd party interfaces for data interpretation. Integration with the BI services and Enterprise Application Integration (EAI) services to third party agents for advanced analysis and dashboard generation
Technical Prcoess ow
There are ve stages depicted in below diagram, stating the lifecycle of the data process in big data platform 1. Data Capture Stage - Fetch the customers E&P data from various well in different phases. Apache Flume can be used for capturing Oil Well log data embedded in a standard formats such as Logical ASCII Standard (LAS) les, Seismic Data les etc. Sqoop can be used for capturing data from RDBMS structured production data. 2. Data Storage & Preparation Stage The massive volume of relevant data is then stored on Hadoop distributed le systems. Hadoop streams can be used for invoking the data preparation, massaging and cleansing scripts. The data preparation jobs can convert the unstructured data to structure format, perform cleansing and store in unied Hive structures. The Data Governance can be carried out by tools such as Oozie and Zookeeper
Fig. 2 : Data process ow in Big Data Enabled Digital Oil Field analytics Source: iGATE Research
Continued on Page 36
CSI Communications | April 2013 | 18 www.csi-india.org
Article
Big Data
Big Data[1] is a large volume of data from various data sources such as social media, web, genomics, cameras, medical records, aerial sensory technologies, and information sensing mobile devices. Big Data includes structured, semi-structured, and unstructured data. This unstructured data contains useful information which can be mined. Since 1980s, per-capital capacity to store information is increased into double the amount for every 40 months. In 2012, statistics says that 2.5 quintillion (2.5 * 218) bytes of data are created per day. Moreover, digital streams that individuals create are growing rapidly. For example, most of the people are using camera on their own. Big Data are of high level volume, high velocity, and high variety of information that needs advanced method to process the Big Data. In addition, conventional software tools are not capable of handling Big Data. So Big Data requires extensive architecture. The following types of data are referred to as big data. Social data Customer feedback forms for Customer Relationship Management (CRM) in Social media sites such as Twitter, Facebook, LinkedIn etc. Machine-generated data Sensor readings, Satellite communication Traditional enterprise data- Employee information, business product, purchase, sales, customer Information, and ledger information. Storages Search Sharing Analysis Visualizations Why Big Data? Big Data is absolutely essential for the following intents: To spot business trends Determine quality of research To prevent diseases To link legal citation To combat crime To determine real time roadway communication system, where the data is created in the order of exa bytes (218). Where it is used? Areas or elds where big data are created: Medicine, Meteorology, Connectomics, Genomics, Complex Physics Simulation, Biological, Environment Research, and Areal Sensory System (remote sensing technologies). Big Science, RFID, Sensor Networks. Astrometry.net project keeps eye on Astrometry group via icker for new photos of the night sky. It analyzes each image and identies the celestial bodies such as stars, galaxies etc. log anytime and anywhere. To assure more security individuals are given with their own login and data stored over the cloud would be encrypted. PHR includes variety of data such as structured, unstructured, and semistructured. In PHR, we propose machine generated data by acquiring the nger print or iris pattern or face of the patient for saving the entire data log of the patient. It uses nger print sensor or Iris scanner or face recognizer for capturing the patient Identication. Finger print or iris pattern or facial expression act as a key for retrieving the data saved in the database Traditional enterprise data includes the entire PHR right from his/her birth with the details of the doctors and their prescription and all records. PHR called as social data which can be made online for online consultation and medicine purchase. Even the lab test reports can be uploaded online. This avoids patient waiting time in the lab for the result report. A copy of the result report will also be sent to the respective consulting doctor for further enquiry. An individual login is provided for patient, doctor, pathologist, pharmacists,etc. , which makes the system more secure.
MapReduce
MapReduce[2] is a programming model for handling complex combination of several tasks and it was published by Google. It is a batch query processor and can run an ad hoc query for whole dataset and get the results in a sensible manner which has to be transformative. It has two steps. 1. Map: Queries are divided into sub queries and allocated to several nodes in the distributed system and processed in parallel. 2. Reduce: Results are assembled and delivered.
Conclusion
Omar Tawakol, CEO, Bluekai has written an article recently. In that article, he has mentioned that More data usually beats better algorithm. But it is very rigid to store and analyze. However, Big Data are used for nding the customer behavior, for identifying the market trends, for increasing the innovations, for retaining the customers, for performing the operations efficiently. Flood of data coming from many sources must be handled using some non-traditional database tools. It provides more market value and systematic for the upcoming generation.
Traits of Big Data Big data differs from other data in 5 dimensions[3] such as volume, velocity, variety, and value. Volume: Machine generated data will be large volume of data. Velocity: Social media websites generates large data but not massive. Rate at which data acquired from the social web sites are increasing rapidly. Variety: Different types of data will be generated when a new sensor and new services. Value: Even the unstructured data has some valuable information. So extracting such information from large volume of data is more considerable. Complexity: Connection and correlation of data which describes more about relationship among the data. Challenges Storing and Maintaining the Big Data is a challenging task. The following challenges need to be faced by the enterprises or media when handling Big Data: Capture Duration
Database
Oracle has introduced the total solution for the scope of enterprise which requires Big Data. Oracle Big Data Appliance[3] is a tool to integrate optimized hardware and extensive software into Oracle Database 11g to endure the Big Data challenges. Example Application: Patient Health Information System on Cloud The Real-time application of Big Data can also be in Patient Health Information System on Cloud[4]. Patient Health Record (PHR) is an emerging technique to store the Patient Heath Information Record and exchange the data over the network, which is stored at the cloud for accessing the data
References
[1] Wikipedia, the free encyclopedia [2] White, Tom, Hadoop: The Denitive Guide. O'Reilly Media, ISBN 978-1-44933877-0. [3] An Oracle whitepaper, Jan 2012 Oracle: Big Data for the enterprise. [4] Scalable and Secure Sharing of Personal Health Records in Cloud Computing using Attribute-based Encryption, M. Li, S. Yu, K. Ren, and W. Lou, Sep 2010 n pp 89- 106.
Article
Jyotiranjan Hota
Associate Professor, School of Management, Krishna Campus, KIIT University, Patia, Bhubaneswar
Introduction
In-Memory Analytics facilitates querying of data from Random Access Memory instead of physical disk. Detailed data can be loaded from multiple sources into the system memory directly. This technique helps in taking faster business decisions. Performance is improved as storage and operations are performed in the memory. The approach of In-Memory Analytics has brought a paradigm shift in storage philosophy. Here, summarized data are stored in RAM. However, In case of databases, data is stored in tables through relationships and interconnection among tables, and other database objects. Similarly, multidimensional cubes are created and data are stored in traditional business intelligence platforms. In case of In-Memory Analytics, the creation of Multidimensional cubes are avoided[1]. As per Gartner, capabilities of in-memory analytics includes faster query and calculation which almost avoids to build aggregate and precalculated cubes. Some myths and facts of in-memory approach are described below (Fig. 1).
In-Memory is just a Hype spread by SAP All major software vendors deliver in memory technology Its new and unproven technology It has been around since 1990s It is solely about running analytics faster Its widely used for transaction and event processing as well Its incremental and nondisruptive Prediction :In-memory will have an industry comparable to web and cloud
Source: Gartner6
Oracle Exalytics
Organizations needs analytics for gaining insight, so as to take correct decision. However, due to budgetary pressure, time
www.csi-india.org
databases. Total cost of ownership of the rms is reduced and business performance is enhanced.
Research Challenges
There are few research issues and challenges of in-memory analytics. In-memory analytics must face the challenge of technology incumbency, particularly in companies where there is heavy dependency on traditional OLAP technology. Many organizations have entire departments built around certain business intelligence platforms, and the threat of any disruptive technology that may signicantly reduce, even eliminate these empires will be met with resistance and skepticism. Enterprise reporting has emerged as a mission-critical function, and once the user community is dependent upon large numbers of reports, one should hesitate before introducing too much change, too fast[1]. As per the IDC Report (2011), traditional method of building and developing computing infrastructure in case of analytics applications are not suitable, when migration to in-memory analytics applications is needed.
sensitivity, and extensive requirement, IT rms usually face challenges to produce actionable analytics. The task even becomes more complex due to involvement of multiple hardware, networking, software, storage vendors, and expensive resources are wasted integrating software and hardware components to generate complete analytics solution. Oracle Exalytics is an optimized system, which provides solution to all business related issues without compromising speed, simplicity, manageability, and intelligence. Oracle Exalitics is built with market leading BI software, in-memory database technology, and industry-standard hardware. Oracle claims that exalytics uses a new interface designed to produce quick result regardless of the query, location, and device types[3].
exact analysis. In few rms ,the need for consolidated reporting and forecasting is required frequently within 10 to 12 weeks .In-Memory analytics is quite t to be used in these circumstances[1] .In the current context, there is a drastic drop in prices of memory and processors. At the same time, multi-core processors are evolving. In-Memory computing has made it possible to perform storage and operations in main memory, where the requirement of hard disk can be avoided. Due to two valid reasons, In-Memory computing is useful. Firstly, the volume of information is growing at an alarming rate. Secondly, immediate responses are needed as quick decisions are needed now in all forward looking organizations. Traditionally, annual and quarterly review reports were taken as the basis for decision making. Past data analysis using data warehousing technology is slowly vanishing. In-Memory computing is supporting event driven systems, which enable decision making in real time. Here data is brought closer to central processing unit. Compared to disk based access, the querying of the data based on in-memory is million times faster. Adoption of 64-bit architecture is a facilitator to in-memory approach as the addressable memory space is increased. Usually midsized companies lack in technical expertise and resources to construct data warehouses, and performance tuning tasks. However, inmemory approach for midsized companies is less cumbersome, easy to administer and set up. IT Infrastructure is not a barrier here, in optimizing business performance. In-memory approach reduces the skill gaps in constructing and consuming analytical applications. The reason for reducing the difficulties is due to avoidance of use of OLAP cubes, which are stored in back end
Table 1 Vendor Dell Fujitsu Fusion IO HP IBM NEC Oracle SAP Kognitio Advizor Solutions Microsoft QlikTech Quantrix Quartet FS SAS Sybase TIBCO
Website http:/ /www.dell.com http:/ /www.fujitsu.com http:/ /www.fusionio.com http:/ /www.hp.com http:/ /www.ibm.com http:/ /www.nec.com http:/ /www.oracle.com http:/ /www.sap.com http:/ /www.kognitio.com http:/ /www.advizorsolutions.com http:/ /www.microsoft.com http:/ /www.qlikview.com http:/ /www.quantrix.com http:/ /www.quartetfs.com http:/ /www.sas.com http:/ /www.sybase.com http:/ /www.tibco.com
Hardware Solution VIS Next Generation Datacenter Platform;PowerEdge R910 PRIMEQUEST 1800 Series; FCRAM; FRAM Fusion IO Flash Memory HP Converged Infrastructure Platform; ProLiant DL 900 Series IBM soildCB Express 5800/A1080a Exalytics In-memory Machine SAP High Speed Analytical Appliance (HANA), SAP In-Memory Computing Kognitio WX2 Analytics Database, WX2 Datawarehouse Appliance, DaaS Cloud
Analytics Solution
Advisor 5.8; Advisor Analyst PowerPivot QlikView DataNaV Active Pivot In-Memory Analytics Adaptive Server Enterprise (ASE) Spotre
[2] Schwenk, H (2010). Accelerating timeto-insight for midsize companies using in-memory analytics available at http:// www2.technologyevaluation.com/ppc/ request/whitepapers/acceleratingtimetoinsight-for-midsize-companiesusing-inmemory-analytics.asp fetched on 1st February 2013. [3] Gligor, G, Teodoru, S (2011). Oracle Exalytics: Engineered for Speed-ofThought Analytics. Database Systems Journal, 2(4), 3-8. [4] Kajeepeta, S (2012). The Ins and Outs of In-Memory Analytics, available at http://www.informationweek.com/ sof tware/business-intelligence/ the-ins-and-outs-of-in-memoryanalytics/240007541 fetched on 29th September 2012. [5] Morriss, H D (2011). Faster, Higher, Stronger: In-Memory Computing Disruption and what SAP HANA means for your Organization, available at download.sap.com fetched on 15th March 2013. [6] Pezzini, M (2011). The Next Generation Architecture: In-Memory Computing, available at http://www.slideshare.net/ SAP_Nederland/the-next-generationarchitecture-inmemory-computingmassimo-pezzini fetched on 25th March 2013. [7] Groth, H (2012). SAP HANA-Strategy and Roadmap, available at http:/ /www. saptour.ch/landingpagesfr/Manager/ uploads/23/32.pdf fetched on 25th March 2013. [8] Chumsantivut, B (2011). SAP HANA Power of In-memory Computing, available at http://www.cisco.com/ w e b / T H /a s s e t s /d o c s /s e m i n a r/ SAP_HANA_Power_of_In_Memory_ Computing.pdf fetched on 25th March 2013. [9] Dale, S (2011). Getting real-time results with in-memory technology, available at http:/ /enterpriseinnovation.net/ article/getting-real-time-resultsmemory-technology fetched on 25th n March 2013.
being used in variety of applications like risk management, inventory forecasting, protability analysis, fraud detection, algorithmic trading, and areas like sales incentive promotion management. Refactoring existing applications in-memory to utilize the approaches of in-memory can result in better scalability and transactional application performance, lower latency application messaging, drastically faster batch execution, and faster response time in analytical applications. In year 2012 and 2013, cost and availability of memory intensive hardware platforms reach tipping points. So the in-memory approach will enter the mainstream.
References
[1] Baldwin, T (2008). Don\t fold your cubes Just Yet But In-Memory Analytics is beginning to Mature, available at http://www.tagonline.org/ articles.php?id=298 accessed on 24th October 2012.
Prof. Hota is an Associate Professor and Area Chairperson of Information Systems wing at KIIT School of management, Bhubaneswar. He is a BE in Computer Science, from NIT Rourkela and PGDBM from Xavier Institute of Management, Bhubaneswar. He teaches Data mining, Business intelligence, Analytics, and core modules of SAP ECC 6.0 like SD, MM, FI-CO, HCM, and PP Modules in view and conguration modes. His research interest lies in banking technologies, analytics, and ERP. He has published several papers in many Journals and Conferences in India and abroad. Author can be reached at jyotiranjan_h@yahoo.com .
www.csi-india.org
CRISC
Gold Winner for Best t n Professional Certication m Program
8 JUNE 2013
Article
Risk Response
The probability of the occurrence of risk may be difficult to predict, but one can never assume it to be zero. Sooner or later the hypothetical risk scenario may actually materialize. It is desirable to be adequately prepared with risk response. The purpose of dening a risk response is to ensure that the residual risk is within the limit of the risk appetite and tolerance of the enterprise. A risk professional will have to clearly dene the risk response options. Every response will have to be evaluated with cost/benet analysis and weighed against a number of parameters including the cost of response to reduce the risk within tolerance level, importance of risk, capability to implement response, effectiveness of response as well as efficiency of response. The available risk response options are to (a) avoid the risk, (b) reduce/mitigate the risk, (c) share or transfer the risk and lastly, (d) accept the risk. It may not be an easy job to decide on an appropriate option. Although there are major risks in electronic commerce transactions, avoiding e-commerce is not really an option today. A thorough cost/ benet analysis has to be done among the remaining three options before taking a decision. This will require building up a business case to justify the selection of response. The risk professional will have to be very familiar with organizational risk management policies, portfolios, investment and value management, exception management, parameters for risk response selection, risk appetite and tolerance, and the concept of residual risk.
Risk Monitoring
A pre-planned risk response is essential to effectively and efficiently deal with a risk. If there is a good risk monitoring
www.csi-india.org
process implemented to keep a watch on various risks and sound an alarm as soon as some risk parameters cross the risk threshold, it will denitely save much of the efforts that will go in responding to the risk. Developing these risk indicators will be a major challenge. There may be literally hundreds or risk indicators such as logs, alarms, and reports. A risk professional will have to closely work with senior management and business leaders to determine which risk indicators will be monitored on a regular basis and be recognized as Key Risk Indicators (KRIs). The KRIs should be selected based on the following factors: Reliability, i.e. they will every time sound an alarm without fail Sensitivity, i.e. the alarm will be sounded only when a certain threshold is reached Impact, i.e. the KRIs will be selected for areas which will have high business impact Effort, i.e. the preferred KRIs will be those which are easier to measure The risk professional should be familiar with various risk monitoring sources. The information for risk monitoring could be obtained from suppliers or vendors of hardware, software, applications in terms of various updates, anti-malware vendors, logs of devices, CERT alerts, newspapers, blogs, and technical reports published by information security research organizations. This means that the professional will have to constantly update the knowledge.
include the feasibility study, requirements study, requirements denition, detailed design, programming, testing, installation, and post-implementation reviews. The business risks could be the likelihood that the new system may not meet the users business needs, requirements, and expectations. The project risks could be that the project activities to design and develop the system exceed the limits of the nancial resources set aside for the project. As a result, the project may be completed late, if ever.
was established in April 2010. While CRISC is designed for risk professionals with at least three years of experience, more than 1,200 CIOs, CISOs, and chief compliance, risk and privacy officers have also chosen to pursue the designation. CRISC is the result of signicant market demand for a credential that recognizes experienced risk and control professionals. This demand will only accelerate as stakeholders demand better corporate governance and business performance, and more secure infrastructures. If you have real-world IT controls and risk experience, I strongly encourage you to pursue the CRISC certication. Becoming CRISC certied provides an additional level of assurance that you have the necessary skills and experience to get the job done. It also enters you into a group of professionals with common interests and abilities. Networking with my fellow CRISCs and ISACA members has been an extremely rewarding experience. I encourage you take advantage of the opportunities certication provides. Shawna Flanders, CISA, CISM, CRISC, process engineer at PSCU, USA In India, we are making rapid progress in the adoption of information technology. Organizations are well aware that they should not take undue risks to achieve their ambitious goals, and should build appropriate IT controls to manage the risks. This has prompted rapid acceptance of CRISC certication in India, and created new job and promotion opportunities for CRISC certied professionals. Avinash Kadam, CISA, CISM, CGEIT, CRISC, CISSP, CSSLP, GSEC, GCIH, CBCP, MBCI, PMP, CCSK, is an advisor to ISACAs India Task Force. ISACA is a global association for IT assurance, security, risk, and governance professionals with more than 100,000 members worldwide and more than 6,000 in India. The nonprot, independent ISACA developed the COBIT framework for governance and management of IT, and offers the CISA, CISM, CGEIT, and CRISC certications. Opinions expressed in the blog are Kadams personal opinions and do not necessarily reect the views of ISACA (www.isaca.org).He can be contacted via n e-mail avinash@awkadam.com.
Practitioner Workbench
Programming.Tips ()
The range function is also very useful to python users .lrange creates a list of numbers in a specied range range ([start,] stop [, step]) -> list of integers When step is given, it species the increment (or decrement).
>>>range (7) [0, 1, 2, 3, 4,5,6] >>>range (7, 12) [7, 8, 9,10,11] >>>range (0, 12, 2) [0, 2, 4, 6, 8,10]
This range function is heavily used in for loop. For example, if you want to print every third element in a list?
for i in range(0, len(array), 2): printarray[i]
It is worthy to mention here that the above chart does not cover all the packages or libraries of Pythons. It only tells about a very small subset of available toolkits or packages using Python. Some special kind of features, which make it so robust are given below: Declaring multiple variables of different types simultaneously in a single line:
>>>x, y, z = A, 2, 5.6 >>>x A >>>y 2 >>>z 5.6
Supports Complex variables Examples:3+4j, 3.0+4.0j, 2J #Must end in j or J Strings are repeated with the * sign: >>>xyz*3
xyzxyzxyz
Python
also
supports
negative
indexes.
For
example,
stringExample[-1] means extract the rst element of stringExample from the end Apart from string, Pyton support List which is denoted by [],
and can hold numbers, strings, nested sublists, or nothing. The list indexing works just like string indexing. For Example
List1 = [0,1,2,3], List2 = [zero, one], List3 = [0,1,[2,3],three,[four,one]], List4 = []
It is possible to return multiple values from a function. Even the documentation of function is given with some corresponding syntax
#Function definition def remove_duplicated(arg_referents, arg_ conditions): This function removes duplicates from two list arg_referents and arg_conditions
It is possible to append, extend, insert, and remove data using the following syntaxes
a. list.append(x) b. list.extend(L) c. list.insert(i,x) d. list.remove()
It is possible to sort a list, count the number of elements and reversing a list using the following syntaxes
a. list.count(x) b. list.sort() c. list.reverse()
N Nibaran Das received his B.Tech degree in Computer Science and Technology from Kalyani Govt. Engineering College u under Kalyani University, in 2003. He received his Mastes in Computer Science and Engineering (M.C.S.E) and Ph. D. degree from Jadavpur University, in 2005, and 2012 respectively. He joined J.U. as a lecturer in 2006. His areas o of current research interest are OCR of handwritten text, Bengali fonts, and image processing. He has been an editor o of Bengali monthly magazine Computer Jagat since 2005.
Practitioner Workbench
Programming.Learn("R")
R- StaR of Statisticians
http:/ /www.r-project.org/ If your requirements are to manipulate, model or visualize a huge set of statistical data, an arguably best choice of programming environment is R!!! R is a follower of Scheme and S Plus; a functional programming language (S language) developed at Bell Labs by John Chambers and team, and is the most widely used one in the area of statistical computing. R was initially developed in 1993 by Robert Gentleman and Ross Ihaka at the Statistics Department of the University of Auckland, New Zealand, and was later progressed as a result of the collaborative effort with contributions from all over the world.R has got its name by debiting the rst letters of its initial developers. R has an intuitive and easy syntax for even a beginner, who has basic programming experience. Like other programming languages, R also has the standard control structures and can be accessed from the languages such as Python, Perl, Ruby etc. Also commercial software such as Mathematica, MATLAB, and Oracle support R in a pretty good way. R recommends command line interface (CLI), which is best suitable for programmers. However, for a beginner to start with, GUI based code editors, and IDE are useful. It provides a variety of functionality like syntax highlighting, code completion, and auto code indentation, which eases the job. Rstudio, Vim-R-Tmux, Notepad++, RKWard, R Commander are some of them. Let us have a look at R interface and R programming environment. When you launch R, the R console will appear with some basic information regarding R, within the R Gui. The console resumes with a prompt with a > symbol. This shows that the interpreter is ready and is waiting for your R commands. We can input commands (which are referred to as expressions) in R through the R console. R programming language has become an important platform for statisticians to work with. R, being.. an open source platform, there are n-number of freely available packages using which, you can, not only do your serious statistical analysis, but also can be used as a analysis platform for the problems in the elds, such as Bioinformatics, Financial Market Analysis, Pharmacokinetics, Natural language processing etc. We can explore more about R programming in the next issue. Have a great time ahead.
Ross Ihaka
Robert Gentleman
R is an interactive, objectoriented language; designed by statisticians for the purpose of statistical computing. It is free and open source, and is available under GNU - General Public License version 2. It runs on most of UNIX platform, Windows, and MacOS. Different versions of R are available at Comprehensive R Archive Network (CRAN), which is a repository for R code and documentation. CRAN also provides source codes, new features, and bug xes etc. Currently there are 4415 packages for R. Now R has become a favorite language for data analysis and statistical computing for both corporates and academia. R is also being used for handling and analyzing large datasets obtained from supercomputing applications and to create high quality visualization via different types of plots like line plots, contour plots, and interactive 3D plot.
CIO Perspective
Introduction
One of the computing areas that is attracting a lot of attention is 'Big Data'. What exactly is Big Data? As per Wikipedia, 'Big Data is a collection of data sets so large and complex that it becomes difficult to process using on-hand database management tools or traditional data processing applications. The challenges include capture, curation, storage, search, sharing, analysis, and visualization'. As more and more activities are being carried out on the Internet by people and enterprises, the amount of data generated for the various activities is rising each day. As per one statistic, as of 2012, 2.5 quintillion (2.5 x 1018) bytes of data were created. While on a personal level users are facing challenges with large volumes of data, the challenge for enterprises is monumental. Enterprises are struggling to derive meaningful value from humongous
www.csi-india.org
data. Such technologies range from simple operational dashboards based on conventional Database Management Systems to advanced techniques like InMemory Real-time Data Analytics.
Technology Overview
The concept of Data Stream Processing is built using the computer programming paradigm based on the 'Single Instruction Multiple Data (SIMD)' parallel programming design pattern. In particular, this paradigm utilizes the concept of 'Pipeline Parallel Processing'. To help understand the concept, refer Fig. 2. In most cases, enterprises continuously receive data that needs to be processed. This data can be viewed as a 'Stream of Data over Time'. For a realtime response, this stream of data needs to be processed, rened and acted upon in real-time. The concept of Data Stream Processing enables real-time processing of such continuous data streams.
process continuous ow of data. A stream is a sequence of time-stamped data records called 'tuples'. A tuple is similar to a row in a database table. As illustrated in Fig. 3, the tuples in a stream have a schema, which denes each eld's name, position, data type, and size. A few examples of a data stream include nancial trading data and sensor data. Window Processing A 'Window'
data stream into a window. When the window is full, the oldest data tuple is pushed out. Domain Specic Language To help ease the task of describing how the incoming data is to be processed, data stream engines typically provide an expressive language - a Domain Specic Language (DSL) - that allows enterprises to dene complex relationships among the data items. As depicted in Fig. 5, data processing rules (queries) can be dened using the DSL. Once dened, the rules act as continuous queries that are deployed once and continuously process the data items streaming by, producing results. Most DSLs are dened to be similar to SQL so as to leverage developer familiarity thereby increasing productivity and reducing maintenance efforts.
The concept differs from conventional data processing frameworks and solutions in several ways, as below: Data streams are usually unbounded. No assumption can be made on data arrival order. Size and time constraints make it difficult to store and process data stream elements after their arrival.
Key Characteristics
Data stream processing engines have the following characteristics: Data Stream Management The engine needs to have capability to
is one of the key concepts of data stream processing. It enables limiting the portion of an input stream from which elements can be selected. While processing a stream of data, it is necessary to dene portions of the input ows that have to be considered while executing the processing rules. Each window contains, at any given time, a subset of the tuples streaming by. Dening such windows enables a query to identify the nite set of tuples (from an otherwise innite stream) over which the processing rules would be applied. As illustrated, Fig. 4 describes how a window is applied on a data stream for say a Withdrawal transaction. The size of the window is ve. The engine stores all arriving withdrawal
To help enterprises process and analyze high velocity data, we have dened the 'iGATE Analysis and Intelligence in Real-timeTM' approach. The approach is built on concepts of Data Stream Processing and Complex Event Processing and has evolved from our experience of implementations across domains. In the following sections we would like to share our experiences in creating solutions for domains like Manufacturing Industry, Smart Grid and Oil and Natural Gas.
dashboards. A dashboard for monitoring plant performance is depicted in Fig. 7. Some of the features of the solution are: Continuous queries to analyze and transform streams of data in real-time. Integration of business intelligence across different applications in real-time. Scalable platform capable of processing vast volumes of real-time data.
business need. This continuous stream of data can be used to provide consumers with an accurate and up-to-date picture of their energy consumption. As depicted in Fig. 8, the high velocity data was processed using a data stream processing solution. The solution processes energy consumption data and presents the same using real-time dashboards. The solution also provides real-time notications and allows aggregated data to be persisted to a data warehouse for further analysis. As illustrated in Fig. 9, the consumption monitor displays real-time consumption data by itself or juxtaposed with historic data. Some of the key features on the solution are: Provides meaningful insight into realtime data for improved customer experience. Improved performance by using inmemory processing. Hybrid approach of in-memory processing of high volume real-time data to provide immediate useful feedback and persisting aggregated data in warehouse for future data analysis.
www.csi-india.org
other upstream processing scenarios. Multiple scenarios, including improving well performance and generating alerts for disaster management can leverage in-memory architecture based on data stream processing. As shown in Fig. 11, the real-time dashboard related to well drilling allows the upstream professional to benchmark and monitor crucial operational metrics with real time data. Some of the key features on the solution are: Multiple real-time data streams that are aggregated on-the-y. Allows enrichment of real-time data with data from the static operational data store.
Benets
The iGATE AIR approach helps generate results as soon as the input data becomes available, delivering business intelligence continuously and in real-time, which can be consumed by applications, services, and users throughout the organization. The benets of the approach can be broadly classied into two categories, namely business benets and technology benets, some of which are given below:
that support optimal short-term and longterm decision making. Its highly trained professionals need integrated views, oftentimes related to a particular process or production event. Despite complexity, engineers, and operators must quickly
identify signicant process events; assess their relevant parameters and take suitable actions. Upstream professionals nd operational insights useful in exploration, drilling and completion, production and
Business Benets
Smarter integration of real-time business intelligence across the organization. Improved business agility, business innovation, and business continuation. Reduction in development time and cost by using standards-based SQL. Reduced storage cost, as data is not required to be persisted before it can be analyzed. Allows real-time data collection, transformation, aggregation and reporting. Lower latency, as data can be analyzed in-memory, before it is persisted to the storage medium. Data independence, that is, logical/ physical separation, leading to loosely coupled applications that need lesser tuning and are more exible. Can be integrated with multiple stream processing solutions like StreamBase, SQLstream, and Esper, to name a few.
Technology Benets
Conclusion
The concepts of data stream processing leverage the performance improvements on the hardware side, particularly the developments in RAM technology, allowing for in-memory processing. In turn, in-memory processing allows enterprises to perform data analysis in
real-time, enabling real time visibility into the day-to-day operations. It is important to note that Big Data is not only about processing, cleaning or churning high velocity, high volume data. It is about deriving some relevant, meaningful insights from the same. Solutions built using Data Stream Processing concepts
can be used effectively to analyze high velocity operational data and extracting business insights in real time. As described in this document, we have used the concepts of data stream processing to build solutions for operational insights across multiple domains. Operational insights can be leveraged to improve the speed of visibility into key operational metrics, thereby helping improve business agility. The iGATE Analysis and Intelligence in Realtime approach aims to augment existing data warehouse, and business intelligence solutions, to enable real-time data processing, rather than seeking to replace them. This allows enterprises to effectively use their existing solutions, but add the capability of real-time data processing, helping them respond to events much n faster and in an effective manner.
Bipin Patwardhan is a Technical Architect with more than 15 years of experience in the IT industry. At iGATE, he is leading the High Performance Computing CoE. The CoE builds capabilities around technologies that help delivery high performance for enterprise applications. Presently, the CoE covers areas like Parallel Programming, GPU Programming, Grid Computing, Real-time Analysis and In-Memory Computing.
About the Authors
Sanghamitra Mitra is a Technical Architect from R&I (Research & Innovation) group in IGATE. She has around 15 years of experience. She has worked on various projects related to Enterprise Applications as well as Enterprise Application Integration with international clients across multiple domains. Currently, her primary focus is on hands-on evaluation of emerging technologies in High Performance Computing area including Parallel Computing, Real-time Intelligence. She is responsible for institutionalizing these technologies across the organization and leveraging them to build innovative solutions to solve business problems.
www.csi-india.org
Security Corner
Information Security
According to CERT-In (Computer Emergency Response Team India), till October an estimated 14,392 websites in the country were hacked in 2012.
schematics for toys or automobile designs. Their motivation can be nancial gain, a competitors advantage in the marketplace, sabotage of a rival nations essential infrastructure, or even just revenge. APTs start by identifying vulnerabilities that are unique to your employees and infrastructure. And since they are precisely targeted, surreptitious, and leverage advanced malware and zeroday (unknown) exploits, they can bypass traditional network and host-based security defenses. Cybercriminals are increasing the use of Web-based malware, and employing malicious uniform resource locators (URLs) for only brief periods of time. They use throw-away domain names in just a handful of spear-phishing emails before moving on, enabling them to y under the radar of URL blacklists and reputation analysis technology. Additionally, the report points out, they are blending URLs and attachments in email-based attacks, and reproducing and morphing malware in an automated fashion. These techniques render the use of defenses that rely on known patterns of data almost entirely ineffective. We are in April and year 2013, is already the 'year of the hack'. Even more disturbing is the fact that many attacks are being carried out by state sponsored actors from countries like China, Korea, and Iran. It is imperative to know when a targeted attack is underway, and how to gather evidence to be able to understand its purpose and origin. Leveraging multiple security solutions that use different methods to detect malicious activity for both internal and external threats can enhance your capabilities. Security technology has been evolving, and manufacturers are developing ingenious ways of not only detecting, but stopping, zero-day attacks. Many advanced security monitoring tools work well in conjunction with more traditional defenses, such as rewalls, IDPS, antivirus, gateways, and security information and event-management (SIEM) systems. With the right tools in place and staff and operational support behind them, you can gain the situational awareness and counter intelligence needed to identify an attack, and potentially block or quarantine threats. Even if an attack is successful, the insight gained into how it occurred, what information may have been compromised, and the relative effect of your defenses can be invaluable to recovery efforts, and will help you continuously improve your security posture. Indias Cyber Law i.e. under the section 66F (Cyber Terrorism) of The IT Act, 2000 has enough teeth to ght against such criminals if found. India needs to implement a huge knowledge management system which can be used by its defense forces along with DRDO, NTRO, CERT-in. This knowledge management on APT can help us weed of any successful cyber attacks and can increase our cyber attack preparedness. India needs an holistic approach and view to encounter APT threat as a country, we have cyber security heroes in pockets but for APT we need team of heroes guided with systems and processes to channel their ght against APT.
What is APT?
A common denition of APT is hard to come by, as many vendors, consortiums, and groups put their own twist on the terminology. A commonly accepted explanation of APT refers to it as an advanced and normally clandestine means to gain continual, persistent intelligence on an individual, or group of individuals such as a foreign nation state government. APT is sometimes used to refer to sophisticated hacking attacks and the groups behind them. What does that mean to the Indian citizen, though? Simply put, APT is reconnaissance and investigation of your network, in addition to your infrastructure and your information assets. Its a reference to a sophisticated and dedicated attacker or attackers who are willing to lay low and go very slow in exchange for gathering data about you, your organization and how you operate. For the IT Professional managing an environment, adjusting your current infrastructure and preparing for this threat will require a different mindset and some analytical assessment. According to CERT-In (Computer Emergency Response Team - India), till October an estimated 14,392 websites in the country were hacked in 2012. General acceptance is that social media usage boosts the likelihood of a successful APT attempt. Attackers behind APTs are interested in a broad range of information, and are stealing everything from military defense plans like latest DRDO attacks to
India needs to implement a huge knowledge management system which can be used by its defense forces along with DRDO, NTRO, CERT-in.
Reference
[1] h t t p : // f o c u s . f o r s y t h e . c o m / articles/268/Combating-AdvancedPersistent-Threats n
Security Corner
IT Act 2000
www.csi-india.org
IT.Yesterday()
Biji C L
Department of Computational Biology & Bioinformatics, University of Kerala
Birthday Tribute to the Most Inuential Mind of 20th Century-Claude Elwood Shannon
Claude Elwood Shannon, the father of information age was born on April 30, 1916. C.E. Shannons inuence and inspiration underpins everyday life activities ranging from the effect of cell phones to the popular social networking sites. The most commonly used internet jargons save, store, upload, and download arguably symbolize the revolutionary concept laid out by the most inuential mind of 20th century- Claude E. Shannon. It is really interesting to explore the path through which Shannon travelled that nally helped him to propose his landmark ideas. Shannon was known to be very timid, and led a normal childhood. His mother was the principal of the local Gaylord High School and his father was a business man. Shannons was very by y much inspired p y his grandfather, who was a farmer and an inventor. Shannons childhood hero was Thomas Edison, whom he has a common ancestor. As a young boy, C.E. Shannon was a big fan of Edgar Allan Poes The Gold Bug[1]. The Gold Bug is a detective ction that centers on searching the buried treasures by deciphering the secret message. Shannon was even attracted to solve cryptograms right from his school days. Moreover, C.E. Shannon was very curious to learn things. He was interested to know how various devices like model planes worked. He was even adventurous, as young boy; he tried to contact his friend half a mile away, just by hooking the telegraph machine to a barbed wire fence. Apart from this he has a passion towards Dixieland music too. He had a good musical instruments collection. In 1936, C.E. Shannon graduated from University of Michigan, with specialization in both electrical engineering and mathematics. His interest in Boolean logic started from Michigan. Later, Shannon joined MIT, and his acquaintance with Vannevar Bush, dean of MIT's school of engineering changed his whole life. His mentor was very much inuential for recognizing his milestone work on switching theory. It was supposed to be one of the best Master thesis ever produced. He was awarded Alfred Nobel prize in 1940 for his novel contribution to switching theory, which later became the foundation for modern digital systems. With the advice of his mentor, Shannon did his PhD work in the area of genetics. In 1942, C.E. Shannon joined Bell Telephone Laboratories for full time research. During the Second World War time, he worked in aircraft devices and in cryptography. It was highly signicant that the encryption work build in complex scrambling machine, was used by Franklin Roosevelt and Winston Churchill in order to protect their transatlantic communications during war time. It was in Bell lab; Shannon met his wife Betty, a trained Cryptographer. The ambience of Bell labs and its relaxed atmosphere helped Shannon to integrate all his views, which ended in his famous 1948 landmark paper A Mathematical Theory of Communication; C.E. Shannons revelation about the concept of information helps to envisage many developmental changes in the eld of communication. Information theory is the brain child of this great American genius. Inspired by Hartleys paper, Shannon tried to quantify mysterious concept of information. The fresh insight proposed by Shannon helped to realize that quantity of information has nothing to do with its meaning in common parlance. It was Shannons absolutely incredible thinking that helped to relate surprise and information[2]. In the information era, Bit is the fundamental atom of information. It was Shannon who rst used the word Bit as per the suggestion of J.W. Tukey, by coining the two words Binary and Digit[3]. Supremacy of bit lies in its versatility. Bit forms the language for any communication system irrespective of the fact that the message is text, audio or images. All messages are being translated into two states like On (0) or OFF (1) . Shannon proposed even the limit at which a message can be transmitted from one end to another through channel without loss of information. The abstract concept of information proposed by Shannon forms the foundation of all technological advancements, in the eld of data storage and transmission systems. Interdisciplinary approach of Shannon created a revolutionary change in the eld of digital communication. He has astonishingly diversied interests in many elds like Switching, Cryptography, Computing, Articial Intelligence, and Games. His novel contributions helped to shape the modern digital world. More than that, Shannon was an enthusiastic juggler, amazing unicyclist, loved to design many devices out of curiosity, enjoy playing chess, and most importantly an adorable poet and musician. With his amazing mathematical foundation, Shannon laid down the golden rule for modern information theory. Let me close my tribute note by quoting an extract of Shannons masterpiece poem, which was published by John Horgan in Scientic American[4].
A Rubric on Rubik Cubics Strange imports come from Hungary: Count Dracula, and ZsaZsa G., Now Erno Rubiks Magic Cube; For PhD or country rube. This endish clever engineer; Entrapped the music of the sphere. Its sphere on sphere in all 3DA kinematic symphony! Ta! Ra! Ra! Boom De Ay!
With theorems wrought by Conways eight; Gainst programs writ by Thistlethwait. Can multibillion-neuron brains; Beat multimegabit machines? The thrust of this theistic schismTo ferret out Gods algorism! With great Enthusiasm; Ta! Ra! Ra! Boom De Ay! Mens schemes gang aft agley; Lets cube our life away!
References:
[1] Robert Price (1985): A Conversation With Claude Shannon
[2] Arun K S, Achuthsankar S Nair, "60 years since kpbw wcy xz became more informative than I love you, IEEE Potentials (ISSN: 0278-6648), Vol. 29, Issue 6, Nov.-Dec. 2010, pp. 16-19. [3] 1928. C E Shannon, A Mathematical Theory of Communication, The Bell system, Technical Journal, vol.27, pp.379-423, 623-656, July, October, 1948 [4] http://blogs.scientificamerican.com/cross-check/ 2011/03/28/poetic-masterpiece-of-claude-shannon-fathern of-information-theory-published-for-the-rst-time/
Biji C L completed her Master of Engineering from Anna University. She is currently procuring PhD, in the Department of Computational Biology & Bioinformatics, University of kerala.
References
[1] Analytics Magazine (Nov-Dec 2012 issue) - How Big Data is Changing The Oil and Gas Industry by Adam Farris Senior VP of Business Development for Drillinginfo- Austin, Texas http://www. analytics-magazine.org /novemberdecember-2011/695-how-big-data-ischanging-the-oil-a-gas-industry Apache Hadoop Wiki - http:/ /wiki.apache. org/hadoop Hadoop- The Denitive Guide (Book by Tom White, Published by ORielly - June 2009) http:/ /net.pku.edu.cn/~course/ cs402/2011/book/2009-Book-Hadoop% 20the%20Denitive%20Guide.pdf Apache Hive Wiki - https:/ /cwiki.apache. org/conuence/display/Hive/Home Hive A Petabyte Scale Data Warehouse Using Hadoop a paper by Ashish Thusoo, Joydeep Sen Sarma, Namit Jain, Zheng Shao, Prasad Chakka, Ning Zhang, Suresh Antony, Hao Liu and Raghotham Murthy Facebook Infrastructure Team http:/ /infolab. stanford.edu/~ragho/hive-icde2010.pdf
[2] [3]
[4] [5]
Conclusion
A Hadoop based Big Data Framework with Hive as a central Data warehouse layer is widely used to create dynamic and unied
Ab o u t t h e Au t ho rs
Pramod Taneja, Principal Architect, iGATE - Pramod has 20+ years of IT experience and currently leading the Big Data CoE of Research & Innovation group, iGATE. He has served in various capacities managing and supporting business process-led technology as well as strategic management initiatives. Email - pramod. taneja@igate.com Prashant Wate, Technical Specialist, iGATE - Prashant has more than 13 years of experience in IT and is currently part of the Big Data CoE of Research & Innovation group, iGATE. He has extensive experience in architecting and implementing database solutions including Big Data, data modeling, data migration and database optimization. Email - prashant.wate@igate.com www.csi-india.org
Test your Knowledge on Big Data Solution to the crossword with name of rst all correct solution provider(s) will appear in the next issue. Send your answers to CSI Communications at email address csic@csi-india.org with subject: Crossword Solution - CSIC April 2013
1 4 5 6 7 8 2 3
CLUES
ACROSS
2. 5. 8. 9. 11. 12. 14. 17. 19. 25. 26. 27. 28. 29. 30. 31. 32. Document-oriented databases using a key/value interface rather than SQL (5) A space-efficient probabilistic data structure (5, 6) Unit of measurement for data volume (9) Markup language (3) A data ow language and execution framework for parallel computation (3) Structure of data organization (6) A distributed columnar database (5) One quintillion bytes (7) Type of database system that can make deductions (9) Discovery of meaningful patterns in data (9) A massive volume of both structured and unstructured data (7) Type of database designed to handle workloads whose state is constantly changing (8) Different types of data (7) An ordered list of elements (5) The digit one followed by one hundred zeroes (6) An open-source system for processing real time data streams (5) A paradigm for development of distributed computing applications (6,5)
9 11
10 12 14 15 13
16 17 18 19 21 22 23 25 24 20
DOWN
1. 3. 4. 6. 7. 10. 13. 15.
29
26
27
28 30
31 32
One thousand terabyte (8) Required for data persistence (7) An opensource software framework supporting data-intensive distributed applications (6) Rate at which data acquired (8) An open-source database (7) A programming model to process large volume of data sets (9) Method for an integrated knowledge environment (4) Type of database optimized to store and query data that is related to objects in space (7) Type of database with built-in time aspects (8) Technique to clean up noisy data to make this usable (11) Size of date expressed as (6) A very large number (10) Extremely large databases (4) An in-memory computing platform designed for high-volume transactions (4) An engine for query processing and data warehouse (4) Database, in very large form (4)
Did you know about Map-Reduce algorithm for handling huge data?
MapReduce offers a programming paradigm for massive scalability to handle large data volumes. Users specify a map function that which takes a data set as input and transforms it into another set with key/value pairs. Then there comes a reduce function that does the merging these transformed data sets associated with same key in the key/value pairs. (Source: MapReduce: Simplified Data Processing on Large Clusters by Jeffrey Dean and Sanjay Ghemawat, URL: http://research.google.com/archive/mapreduce.html)
S G M L
12
M O N O G
L
8
L
9
E A
10
C O
M O B I
I P V
A Z
O
11
A B
D E B N
15
W
13 14
S T
R A P H
6 O P A L
17
R
16
L I V
A C
X
19
A I F F
25 23
F M P R E D S C E B
27
B R T F
18 20 21 22
T
24
A P A C E R Y
26
O S T S C
29
U S
P U P
G O K R A
Congratulations to Ananthi Nachimuthu (Dept of Computer Technology, Dr. N.G.P. Arts and Science College, Coimbatore) and Madhu S. Nair (Dept of Computer Science, University of Kerala, Thiruvananthapuram) ALMOST ALL correct answers to March 2013 months crossword.
H I
G T
28
T
32
N A
D S S
M
30
R I P
F
31
M L
T E
R Y
33
S L
A W
34
Ask an Expert
~ Mark Twain
C/C++: Catching array index out of bounds From: Anonymous
In C/C++, presumably there is no array index bound exception coming while dealing with raw arrays. Even index operator for a STL vector cannot detect the array index crossing the specied boundary limits. Code snippet follows. #include <iostream> #include <vector> using namespace std; const int SIZE = 2; int main() { int rawarray[SIZE]; vector<int> v(SIZE); int i; for (i = 0; i <= SIZE+1; i++) { rawarray[i] = i; v[i] = i; } for (i = 0; i <= SIZE+1; i++) { cout << "rawarray[" << i << "] = " << rawarray[i] << endl; cout << "v[" << i << "] = " << v[i] << endl; } return 0; } When I compile and run this program, there is no compilation as well as runtime error. However, it is clear from the code given above that each of the rawarray and the STL vector object v is supposed to contain two elements as per specied size i.e. 2 but when I try to put something as the third or even fourth element, they are being allowed without any warning or error. Heres the output: rawarray[0] v[0] = 0 rawarray[1] v[1] = 1 rawarray[2] v[2] = 2 rawarray[3] v[3] = 3 = 0 = 1 = 2 = 3 is unpredictable. A better alternative would be to dene Array as a C++ template and have own exception dened to catch array index out of bounds exception. Code snippet follows: #include <iostream> #include <string> #include <exception> using namespace std; class MyException:public exception { string ex; public: MyException(const string str= "some exception"):ex(str){} ~MyException(){} const char* what(){return ex.c_str();} }; template <class T> class Array { T * data; int size; public: Array(int s) { data = new T[ size = s]; } virtual ~Array() { if (data) delete [] data; } T& operator [] (int index) { if ((index < 0) || (index >= size)) throw MyException ("Array index out of bounds"); return data[index]; } }; const int SIZE = 2; int main() { try { Array<int> safearray(SIZE); int i; for (i = 0; i <= SIZE+1; i++) { safearray[i] = i; } for (i = 0; i <= SIZE+1; i++) { cout << "safearray[" << i << "] = " << safearray[i] << endl; } } catch (MyException &e) { cerr <<"exception: "<<e.what()<< endl; } return 0; } The output would be as below (when array index boundary is crossed): exception: Array index out of bounds For std::vector, the index operator [] does not check for boundary overow or underow. You could use the member function at e.g. v.at(i) and enclose within try block. vector::at throws an out_of_range exception if the requested index position falls out of specied range. Alternatively you may check v.size() to check if you are crossing the specied boundary or not. n
In C/C++, there is no boundary checking for arrays. Even accessing vector with index operator is also buggy, as rightly pointed out. In reality, accessing array element for rawarray with an index say 5 like rawarray[5] actually means you are accessing an element residing at memory location rawarray + sizeof (int) * 5 or rawarray - sizeof (int) * 5. This means that on a 32-bit machine, where sizeof(int) = 32, this will be at location offset 32 * 5 i.e. 160 bytes either way from starting location of rawarray. If that memory location is within the permissible range of memory locations for user programs, no runtime error would come. And if the memory location falls within restricted memory area (reserved by operating system) would cause a protection violation i.e. program would crash. But, the behavior
Send your questions to CSI Communications with subject line Ask an Expert at email address csic@csi-india.org CSI Communications | April 2013 | 38 www.csi-india.org
Happenings@ICT
H R Mohan
Vice President, CS AVP (Systems), The Hindu, Chennai Email: hrmohan.csi@gmail.com
Europe contributes about 25-30% of IT revenues as against 50% from the US markets. Computer users to spend 1.5 bn hours and $22 bn battling malware. Global enterprises will spend $114 billion to deal with the impact of a malware-induced cyber-attack Microsoft. Mobile value added services (MVAS) to reach $9.5 billion in 2015, from $4.9 billion in 2012 Wipro & IAMAI. Cyber security market may reach $870 mn by 2017IDC. Govt. expects lower revenue of Rs 19,440.67 crore from spectrum sale and other related charges in 2012-13, compared to Rs 58,217 crore estimated. Bharti Airtel leads in consumer complaints on billing, tariff. Govt. plans to take over possession of BlackBerry infrastructure in Mumbai, or legal interception of Internet communication. 2G Scam: JPC unlikely to call Raja as witness. May be asked to submit stand in writing. CBI court summons Sunil Mittal, Ravi Ruia, and Asim Ghosh. Raja accuses Vahanvati of telling untruths against him. 2G players face ne from DoT for shutting services without notice. DoT decision to allow broadband players to offer voice is illegal COAI. New messaging system for NGOs with FCRA (Foreign Contribution Regulation Act) registration. Buying Internet protocol addresses to get cheaper, faster with the launching of National Internet Registry (NIR) in India. Free roaming services likely before October Sibal. 14 million requests to switch mobile operator rejected. CDMA spectrum sale to fetch Rs 3,639 cr while the auction of 2G spectrum for GSM players held in November last year fetched Rs 9,407 cr. DoT rm telcos must own spectrum to offer 3G services. Over 2 cr mobile users loaded with valueadded services they didnt ask for. BSNL, MTNL still to recover Rs 6,215 cr from customers Sibal. Time for electronics goods certication extended till July 3. Govt has received proposals for two semiconductor fabs Sibal India plans U.S.-like information sharing to alert cyber-attacks. Telcos asked to install local server for security audit. Fate of Aakash II tablet still uncertain. Sibal unveils roadmap for IPv6. Plan for complete migration to IPv6 by December 2017. Cost of voice services will move up Aircel. Unied license framework to take a month -Telecom Secretary. Centre to set up 2,000 telecom towers in tribal areas at a cost of Rs. 3,000 cr. Freshers hired by HCL Tech stage protests across country demanding that the company
convert the offers to actual jobs. HCL issued a letter of intent and not a job offer HCL HR head Infosys plans to hire 200 in US. Helios and Matheson IT to hire 1,000. Aptech ties up with NSDC. Aims to train over two million people over 10-years. Google to slash 1,200 Motorola Mobility jobs in US, China and India. Sigma Aldrich, to hire 100. Mahindra Satyam to increase headcount in Australia to 5,000 in two years from 1,600. Chennai-born Sundar Pichai to head Google Android division. Fake job offers swarm Android platform. US to accept H-1B visa applications (with a cap at 65,000) from April 1. H1-B visas could double under Senate plan -- Report Hiring activity in IT sector likely to be muted this year -- Kris Gopalakrishnan. Albion Infotel plans to hire 150 people. Nasscom launches programme to incubate 10,000 start-ups. Makuta VFX to double headcount this year from 60. Engg students prefer IT; Google most wanted employer Nielsen. Tyco plans to double headcount from 850. D Shivakumar, Senior Vice-President (India, Middle East, Asia), Nokia decides to quit. SAP training students to meet innovation needs. TCS ranked No. 1 employer in Europe for 2013.
CSI Report
M. Gnanasekaran
Asst. Manager (Administration), CSI
Prizes were presented by the Chief Guest, Mr. S D Shibulal, CEO and MD of Infosys Ltd. during the 48th CSI Foundation Day Celebrations, held at TIFR on 6th March 2013. The committee consisting of R Jaikumar, S P Mudur, V Prabhakaran, Prof. RK Shyamasundar Convener
Kind Attention: Prospective Contributors of CSI Communications -
K Samudravijaya, R K Shyamasundar, and G Siva Kumar evaluated the proposals, and selected the best dissertations for the award. CSI appreciates the remarkable job done by the committee in a very short span of time. n Mr. MD Agrawal Chairman - Awards Committee
Please note that cover themes of future issues of CSI Communications are as follows May 2013 - Cryptography June 2013 - Social Networking July 2013 - e-Business/ e-Commerce August 2013 - Software Project Management September 2013 - High Performance Computing
The articles and contributions may be submitted in the following categories: Cover Story, Research Front, Technical Trends, and Article. For detailed instructions regarding submission of articles, please refer to CSI Communications March 2013 issue, where Call for Contributions is published on the backside of the front cover page. [Issued on behalf of Editors of CSI Communications]
www.csi-india.org
CSI Report
Prof. Prashant R Nair*, Mr. Ranga Rajagopal** and Dr. Rajveer S Shekhawat***
*National Quiz Coord, **National Student Coordinator, CSI ***National Convenor - CSI Project Contest
CSI Discover Thinking Quiz 2013 3rd National CSI Science and ICT Fun Quiz
which were held at Koneru (Region 5), Pune (Region 6), and Kochi (Region 7). The nals saw the top 2 teams from each region competing for the CSI Discover Thinking National Quiz Championship. In all, over 500 schools and almost 5,000 students participated in the quiz in various rounds. The total prize money was over Rs. 2.0 Lakhs. At the end of a pitched battle, Naveen V and Naveen Unnikrishnan of Bhavans Adarsha Vidyalaya, Kochi bagged the trophy and 1st prize of Rs.25,000/-. They were followed by Amal M. and Sarath Dinesh of St. Thomas Higher Secondary School, Trivandrum who took home Rs. 10,000/-. The third place was secured by D Jeevithiesh and M Prabhat of Narayana IIT Olympiad, Vijayawada. A team from Dnyan Prabodhini, Pune came fourth. The prizes were distributed by Mrs. Chitra Buzruk, Senior General Manager, Persistent Systems, in the presence of Mr. Shekhar Sahasrabudhe, RVP, Region 6, and Mr. Arun Tavildar, Past Chairman, CSI Pune chapter. At the Koneru regionals, CSI Past President, Prof. P Thrimurthy distributed the prizes and encouraged the young students. This event was coordinated by Mr. Ranga Rajagopal, NSC- CSI and supported by Prof. Prashant R Nair, National Quiz Coordinator. The nal was anchored by Mr. Shekhar Sahasrabudhe, RVP 6, Mr. S P Soman, RSC 7, and Ms. Mini Ulanat, National Convenor Skill development coordinated the regional round for region 7 while Mr. Praveen Krishna, CSI SBC of KL University Koneru coordinated the Region 5 regionals. CSI Discover Thinking Quiz had Adobe as the event sponsor with Persistent foundation and KL University sponsoring the regional rounds at Pune and Koneru respectively. The quiz aims to encourage young learners to discover science, and ICT the fun way and hopes to address the declining trend of very few children opting for pure science as a profession. n
All eyes were glued to the huge screen. With each passing question, heads came together, hush-hushing the answer. While a few students sat scratching their heads, others clenched their sts in frustration. Fewer still sat under the comfort of utter ignorance. Soon, it was answer-time. And, at once, the quiet auditorium was engulfed by a cacophony of phenomenal enthusiasm, which set the tone for the event. The National nals of 3rd National CSI Discover Thinking Quiz 2013, a Fun Quiz conducted by CSI on 2 March, 2013 at Millennium National School, Pune was for students of middle school from 6th to 9th standard. The quiz master J Ramanand from IBM India, had the whole audience in raptures. An accomplished quiz master, Ramanand is a BBC Micro Mind winner and also founder of quiz club of Pune. Initially, the rst round of this national quiz was conducted in various CSI chapters during January, 2013. The CSI chapter level rounds were held at Trivandrum, Kochi, Sivakasi, Mysore, Koneru, Nashik, and Solapur. The top quizzing team from each chapter moved onto the regional rounds,
CSI Discover Thinking, the 2nd National level student project competition, is an initiative for CSI student members to share innovative ideas with their peers and experts country-wide. The 2nd edition of this extremely popular event was exclusively sponsored and supported by M/s Adobe Inc. A total of 10 teams had been shortlisted after regional rounds (2 per region) to participate in the national round. The National nal was held at College of Technology and Engineering (CTAE), Udaipur on 16th March 2013. All the projects contained ideas & implementations of very high quality and innovation. The competition was very closely contested and at the end of a daylong session of presentations, the following teams were announced as winners. The teams participating in the event are shown in the photo with judges and organizers. 1st Prize: Sai Chand Upputuri & Alakananda Vempala, Behavioural Biometric Advanced Authentication, K L University, Guntur (A.P.) 2nd Prize: (Two teams were rated 2nd) (a) Surya Mani Sharma, Multifunctional Robotic System, Dronacharya College of Engg, Gurgaon (Haryana) (b) Phagun Singh Baya, Remote Wireless Sensors Analysis
CSI Report
Division IV, CSI - Seminar Report on Recent Trends On Computer Security on 25th March, 2013 @ GITA, Bhubaneswar
On 25th March 2013, the CSI student chapter was inaugurated at GITA, Bhubaneswar. A seminar was conducted by Division IV CSI on, Recent Trends on Computer Security. Prof. (Dr.) Sudarshan Padhi, Director, Institute of Mathematics & Applications, Bhubaneswar & eminent computer scientist, attended the seminar as key note speaker. Prof. Padhi delivered his key note address that, computer security is an over changing issue. 50 years ago computer security was mainly concerned with the physical devices that made up the computer. At this time, these were the high value items that an organization could not afford to lose. Today, computer equipment is inexpensive, compared to the value of the data processed by the computer. Now the high level item is not the machine or computer hardware, but the information that it stores and process is more valuable. This has fundamentally changed the focus of computer security, from what it was in early years. Today, the data stored and processed by computer is more valuable than the hardware. Today, one of the most effective measures is security professionals can take to address attacks on their computer systems and networks, is to ensure that all software is up to date in terms of vendor released patches. Virus and worms are just two types of threats that fall under the general heading malware. The term malware comes from Malicious software , which describes the overall purpose of code that falls into this category of threat. We will discuss in details about Recent trends in computer security & steps to minimize the possibility of attack on a system during todays discussion. Division IV Chairman, CSI Sanjay Mohapatra, participated as Honorable guest for this seminar and spoke about CSI & its student chapter activities. He also discussed on different issues of Computer Security. GITA College Principal, Vice Principal, Dean Academics, HOD CSE Department GITA were present in this seminar and addressed students. Prof. Manoj K Pradhan, Student Branch Coordinator proposed the vote of thanks. There were around 200 student participants, and 20 faculty members present in the seminar. n
The Eastern Regional Convention 2013 on Computing Anywhere, Anyware" was conducted at CV Raman College of Engineering, Bhubaneswar from 25th to 27th Feb 2013. This conference was a joint effort of the CSI student branch, in C.V Raman College of Engineering, the CSI Region IV, and CSI Division IV (Communication). The convention was inaugurated by Prof. Dr. Ganapati Panda, Deputy Director IIT Bhubaneswar. In his inaugural address, Dr. Panda explained how evolution of technology has made computing possible anywhere, using small to micro devices, which can be placed anywhere, be it in our surroundings, in the air, water, buildings, homes, and also on and inside our body. These devices which are sensors with processing power and memory can collect real time data, and process large amount of information for several types of applications, meant to help the society, like predicting water level, controlling irrigation or ood, managing road traffic, etc. The Computer Network of today is a hybrid of LAN/WAN, Mobile Network, and Wireless Networks. Some of the challenges that come with advancement of these technologies are - standardization of communication protocol, managing power to the wireless ad-hoc devices widely dispersed, and supporting software to process large amount of data in parallel. Mr. A Pal, Principal Scientist and Research Head, Innovation Lab, TCS Kolkata, speaking on Grid Computing for Internet-ofthings elaborated the concept of connecting all computing devices
www.csi-india.org
CSI Report
Dr. Dilip Kumar Sharma*, Mr. Sanjay Mohapatra** and Mr. R K Vyas***
*Honorary Secretary- Computer Society of India Mathura Chapter **Chairperson, Div IV, CSI ***Vice President, CSI Region-1
An International Conference on Information Systems and Computer Networks: ISCON-2013, was organized in GLA University, Mathura, on 9-10 March 2013, in technical collaboration with IEEE UP Section and CSI Mathura Chapter, Division: IV & Region-1 . It was cosponsored by Indian Oil Corporation Ltd (IOCL). The Chief Guest of the conference was Prof. S K Koul, Deputy Director (Strategy and Planning), IIT Delhi. Mr. R K Vyas, Vice President, CSI Region-1, Prof. M N Hoda, Director, Bharti Vidyapeeth, Delhi & Regional Students coordinator, CSI Region-1, Prof. S K Gupta, Department of Computer Science and Engineering, IIT Delhi, and other dignitaries were present at the Conference. The General Chair for the Conference was Prof. Krishna Kant, Head, Department of Computer Engineering and Applications, GLA University, Mathura. Prof. S K Koul addressed the conference and highlighted some points on being successful: THINK BIG, WORK TOGETHER AND INNOVATIVELY, AND, GIVE MORE AND TAKE MORE. He also guided the participants on writing technical papers. Prof. S V Raghvan, Vice President, CSI and Scientic Secretary, Office of the Principal Scientic Adviser to the Government of India, New Delhi, also addressed the gathering through his recorded video, in which he highlighted the advancements in the domain of electricals and electronics, networking, and semiconductors. He also talked about the National Knowledge Network (NKN), which is a state-of-the-art multi-gigabit panIndia network, for providing a unied high speed network backbone, for all
Announcement
48th Annual Convention CSI 2013 Brochure Released Visakhapatnam Chapter
Mr. S Ramanathan, Hon. Secretary, CSI and Sri HR Mohan, President elect at a programme organised at Visakhapatnam Steel Plant on 16th Mar 2013. The central committee visited Visakhapatnam to review the facilities available for CSI-2013. Speaking on the occasion, Mr. S Ramanathan expressed condence that the Visakhapatnam Chapter of CSI with the all-round support of the PSU giant Visakhapatnam Steel plant will make the Annual convention the most memorable one in the history of CSI. Sri HR Mohan, President elect, said that the Visakhapatnam Chapter has proved its worth by organising several mega events related to IT very successfully and thats the reason s why they chose Visakhaptanam Chapter to conduct this prestigious National convention for 2013. Chair, Sri CK Chand, Sri P Ramudu, Executive Director (Auto & IT), Vice Chair Sri KVSS Rajeswara Rao, GM(IT), Addl. Vice Chair Sri Suman Das, DGM(IT), Sri Paramata Satyanarayana, Convener of Org.Committee and Sri GN Murthy, ED (Finance) & Chair, Finance Committee, Sri DN Rao, ED (Services)& Chair of Convention Committee and Dr. S R Gollapudi, Convener, Advisory Committee were present this occasion.
CSI Annual National convention CSI 2013 is being organised by the Visakhapatnam Chapter in association with Visakhapatnam Steel Plant during 13th -15th Dec 2013 at Hotel Novotel, Visakhapatnam. The theme of the Annual Convention is IT FOR EXCELLENCE . It will be held in Visakhapatnam for the 1st time in the history of CSI, since its inception 48 years ago in India. To mark the beginning of the arrangements of the annual convention, a colourfully designed brochure giving various details of the Convention was released by Sri Umeshchandra, Director (Operations) and past Chairman, Visakhapatnam Chapter in the august presence of
CSI Report
VI/Runner-Up
II/Special Mention
Mr Bipin Mehta, Dr. R. Rudramoorthy, Mr. Ranga Raj Gopal and Award Winners
chapter. An announcement of YITP awards was sent to all the chapters, corporate members, institution members and IT companies including announcements in CSI Communications and hosted on CSI website. This resulted into getting many nominations. After short listing the nominations, 40 teams comprising of professionals from IT companies, technical institutes, entrepreneurs and researchers participated at Regional Level. CSI YITP| Awards maintains absolute transparency in an objective and merit based selection process. This year the evaluation process included 2-tier selection process to select the Winners, Runners-up & Special Mention. The regional round was conducted at Kolkata, Ahmedabad, Bhilai, Bangalore, Nashik and Chennai Chapter. With the support of Regional Vice Presidents and Regional YITP Conveners, the regional round completion was successfully conducted on the above six regions. Details of the regional round competition can be viewed on the CSI website under CSI News section. From each region, Winner and Runner up teams were invited for nal round of competition on 6th March, 2013 at PSG College of Technology, Coimbatore. Total 11 teams presented their projects to the selection committee members. The selection committee members were Dr. Subramaniam, Past Chairman CSI; Mr. John Milton, Robert Bosch; Ms. Pandi Selvi, Robert Bosch; Mr. Sebastian Christopher, CTS and Mr. Isai Amudan, CTS. Mr Bipin Mehta and Mr. Ranga Raj Gopal coordinated the nal round which was supported by Mr. N Valliappan, Secretary, CSI Coimbatore Chapter. The most outstanding technology project of any kind, completed during the year 2011-12 where project duration could be of 2-3 years from the start date, within an organisation were judged. The selection committee considered many factors to judge each project like criticality of IT usage, improvement of customer service, innovation, quality of management and impact on organization and society. It was a challenge to selection committee to decide on the winners. The selection committee unanimously declared the Winners, Runner-up and Special Mention Award Winner as under: The results of the National Round were declared and awards were presented on 6th March, 2013 on the auspicious occasion of 48th CSI Foundation Day at PSG College of Technology. The chief guest for the award function was Dr. R Rudramoorthy, Principal, PSG College of Technology, who inaugurated the contest. In National Round, the winner received Rs. 50,000, a trophy, and a certicate. The runner-up received Rs. 25,000, a trophy, and a certicate, while the team that received special recognition got Rs. 15,000. The contest aimed at involving young IT professionals in the quest for innovation in IT and also providing them an opportunity to demonstrate their knowledge, professional prowess and excellence in n their profession.
www.csi-india.org
CSI Report
Dr PVS Rao
Fellow and Past President of CSI
Opening Remarks
At the outset, Panel Moderator Prof PVS Rao reminded audience that CSI is actually a year older than commonly believed; it is successor to the All India Computer Users Group (AICUG) which was formally started in Faridabad near New Delhi in 1964, (few days after sad demise of Pandit Jawaharlal Nehru) and renamed itself as the Computer Society of India one year later, in 1965. He paid homage to Major General A Balasubrahmanian, Late Prof Bishwajit Nag and Mr. SR Thakur, who along with himself and a few others, started the AICUG in 1964. After welcoming and introducing the panellists, Prof. Rao stated that topic being IT Education and Research, focus will be on leveraging Indias progress in IT for accelerating the pace of National Development and in development of human resources. Coverage would include IT Education itself as well as use of IT in Education. IT research would necessarily include applied aspects (such as software engineering, computer aided engineering activity and so on), which facilitate and catalyse development of IT. It would also necessarily include research as an end in itself (e.g. Theoretical Computer Science). A question to be addressed is as to how best our competence can be leveraged to help in economic development, increasing exports and growing national wealth. Speaking about education in general, Dr FC Kohli said, going by population, India should have three to four times as many bright students as there are in the USA. On the other hand, the annual output is only seven or eight hundred Ph Ds, a number that does not even meet the (teaching) faculty requirements of the various academic institutions already in the country, let alone the numbers needed for research and innovation. This gap needs to be bridged; about 50 colleges have been identied in the country, which can, with proper inputs, bright students and trained faculty, be expected to produce up to 35000 world class graduates annually; of these, 6000 will go on to become Ph. Ds (as against the current output of only 800). Prof. DN Phatak emphasised that need of the day is not merely to ensure that IT training happens on a scale that matches very large numbers of graduates needed; it is most important to provide high quality education. IIT Mumbai is addressing this by training teachers in thousands in a tiered structure so that they can in turn provide quality training to freshers. The trainee teachers are grouped at 40 to 50 widely distributed centers. IIT beams courses covering full range of subjects on-line to these centers in the mornings. Pre-trained course coordinators are available at each of the individual centers for interaction with trainee teachers; they also run tutorials and practical sessions in the afternoons. Prof. SP Mudur spoke about qualitative changes that have occurred over the three decades that he has been in teaching. He cited teacher evaluation system prevalent in Canadian Universities, which lets the student assess the competence of their teachers. In this process, younger teachers are often graded higher than experienced seniors, mainly because older faculty nd it difficult to keep up with changes that are happening. Earlier, student-teacher interaction was restricted to only face-to-face interaction but has been greatly enriched through social networking (U-tube, Face book, Twitter, blogging and so on). Massively On-line Open Courses (MOOCs) such as those offered by MIT are easily available from many reputed universities to students worldwide. Soon, it might be possible to take such courses from multiple universities even for credits. Blended learning (a combination of Face to face as well as on line learning) will become pervasive and important. Scaling will happen as large numbers of students are attracted by high reputation of institutions offering MOOCs. In closing, he mentioned that for the next few years at least, the job situation will continue to be very good for students specialising in Science, Technology, Engineering and Management (STEM). Talking about Industry-Academia interaction, Dr Nirmal Jain emphasised that there are multiple ways of learning and interaction between the two. Often, there is a disconnect between material taught in the Universities on the one hand and what industry really needs on the other.
Only constant interaction between educational institutions and industry can bring about a better match between course curricula and industry requirements. Fortunately, pressures of competition are strongly motivating Industry to interact closely with academic institutions in the hope of gaining a competitive edge by leveraging on the innovations that happen there. It is up to the industry to make many more such collaborations happen.
Intra-panel interaction
During subsequent interaction between panellists, Dr Kohli pointed out that even in todays context of ever increasing prevalence of social networking, faceto-face person-to-person interaction continues to be very crucial. Agreeing, Professor Phatak said the idea is to start (innovative modern methods) in a small way initially and scale up as the process succeeds and proves itself. Dr Jain remarked that as we grow older, rather than just do our jobs, all of us are becoming increasingly interested in and involved with aspects relating to teaching and training; this highlights how important these issues are.
CSI News
From CSI Chapters
Please check detailed news at: http://www.csi-india.org/web/guest/csic-chapters-sbs-news SPEAKER(S) GURGAON (REGION I)
Mr. Vivek Varshney, Mr. R K Vyas, Prof. M N Hoda, Prof. D K Lobiyal, Prof. S K Muttoo and Prof. Jitender Kumar 2 March 2013: CSI Regional level Student Project Contest-2013 The contest aimed at involving students in IT innovation and to provide them an opportunity to demonstrate their projects with strong social relevance. First prize -Mr. Surya Mani Sharma from DCE, Gurgaon project Multifunctional Robotic System. Second prize for Wear Your World to Ms. Monica Bansal and Mr. Deepak Kumar. The project Android Application at third position demonstrated by Mr. Shashank Sharma and Mr. Paras from BVICAM, New Delhi.
KANPUR (REGION I)
Dr. H C Karnick, Dr. Brijendra Singh, Dr. Phalguni Gupta, Dr. Alok Tiwari and Dr. Raghuraj Singh 9 March 2013: National Seminar on Issues and Challenges of Computer Science & Engineering as a Discipline Seminar was jointly organized with Dept of Computer Science & Engineering, Harcourt Butler Technological Institute. A souvenir containing abstracts of invited lectures, expert views of academicians and articles on the seminar theme was released on this occasion. CSI Kanpur Chapter website http:// www.csi-kanpur.org was launched during the seminar.
Guests while releasing the Souvenir
LUCKNOW (REGION I)
Mr. Amit Khanna and Prof. Bharat Bhaskar 15 March 2013: Technical Session on nComputing Thel session was organized at NIEIT, Lucknow centre in association with nComputing and M/s M Intergraph. It was attended by more than 50 participants. During the presentation Mr. Amit Khanna explained the benets, usage and other related details of the product. Prof. Bharat Bhaskar, IIM Lucknow & Chairman of CSI Lucknow Chapter introduced the session.
HYDERABAD (REGION V)
Dr. Pratap Reddy Sir 2 March 2013: Event titled CHALLENGE EXPO-13 Participants demonstrated Project Exhibits and presented Posters during this event. Around 70 projects were exhibited and 10 posters were presented. The event is organized under the guidance of Dr. Pratap Reddy, who presided over as Chief Guest and Judge. Winners were given cash prizes and participation certicates. Details of the event can be found on the url: http://www.dprec.ac.in/challengeexpo13.html along with process of conducting the event and photographs.
CSI CS I Co Comm Communications mmun mm unic un ic cat atio ions ns s | Ap April A r l 2013 ri 201 13 | 46 6
www. ww w.cs w. csics i-in indi in dia di a.or a.or org g www.csi-india.org
Mr. G Santosh Kumar, Mr. Ganta Srinath Reddy, and Mr. Krishna Vattipalli
1-2 February 2013: Southern Regional Conference on Innovative Technologies (SRCIT) 2012-13 Mr. Santosh Kumar covered various areas of hacking, respective counter measures and preparedness. He covered topics such as ethical hacking, major vulnerabilities, basic protection mechanism and penetration testing techniques and importance of its ndings. Mr. Reddy gave introduction to Android programming and deployment with hands on demonstration. Mr. Vattipalli explained deployment and related development on Android and Googles cloud platform GAE (Google App. Engine).
Inaugural Program for SRCIT VIZAG-2013
(L to R:) Principal Gandhe, Principal Tate, Mr. Chandrashekhar Sahasrabuddhe, Mr. Avinash Shirode, Hon. Shivajirao Patil, Mr. Ashok Kataria, Ms. Mohini Patil, Principal Prashant Patil and Shri Shrikant Karode
CSI C CS I Co Comm Communications mmun mm unic un icat atio at i ns | Ap io April Apri ril ri l 2013 2013 | 47
25 January 2013: Seminar on Cloud Computing - Industry Case Studies Mr. Gulabani started with cloud computing basics and various services provided by different cloud providers. Later he shared various cloud computing case studies using Amazon Web Services (AWS) cloud, Redhat Openshift Cloud, Google App Engine and Tumblr. He gave live demo and explained technical architecture of cloud application development using Eclipse and cloud APIs. He also provided innovative project ideas using Google wallet, Geo-based social networking/taxi service, ght back application and mobile e-learning among others.
Mr. Sunil Gulabani, IndiaNic Infotech Pvt. Ltd. shared his industry experience on cloud computing during the seminar
16 February 2013: Lecture on Current and Emerging Trends in ICT, Employment Opportunities and Benets of Professional Society Membership Mr. Sonawala emphasized on information sharing for increasing knowledge. He advised students to use technology for betterment instead of misusing it. He mentioned that technology changes rapidly and hence students should try to achieve expertise in applications and not in tools. He also made students aware about their social responsibilities. Certicates and trophies were awarded to students for their outstanding performance in academics and also for the best System Development Projects.
Mrs. Hemal Desai, Shri Hemant Sonawala and Prof. Bipin Mehta
CSI CS I Co Comm Communications mmun mm unic un ic cat atio ions ns s | Ap April A r l 2013 ri 201 13 | 48 8
www. ww w.cs w. csics i-in indi in dia di a.or a.or org g www.csi-india.org
SPEAKER(S)
Mr. Arunkumar and M Khannur
27 February 2013: One day workshop on Recent Trends in Embedded Systems & Soft Computing Techniques" Dr. Valli Kumariexplained various methodologies and modalities present in the eld of Image Processing. She provided detailed insight for narrowing down the gap between low level image features and human interpretation of the image. She also explained basic concepts of Soft Computing techniques for Image Processing applications.
Honoring the Guest Dr. Valli Kumari
Please send your event news to csic@csi-india.org . Low resolution photos and news without gist will not be published. Please send only 1 photo per event, not more. Kindly note that news received on or before 20th of a month will only be considered for publishing in the CSIC of the following month.
CSI C CS I Co Comm Communications mmun mm unic un i at ic atio i ns | Ap io April Apri ril ri l 2013 2013 | 49
SPEAKER(S)
Mr. Amit Grover and Mr. Siddharth Goyal
Certicate Distribution
JRE group of Institutions, Greater Noida JRE-School of Engineering inaugurated CSI-Student Branch and JRE SOE Project Center on 16th Feb, 2013. Technical talk on RoCK-BEE: Robotics Competition Knowledge Based Education in Engineering by Prof Saha and lecture on Soft Computing and Its Applications by Prof. M MSufyan Beg were organized on this occasion.
REGION V
Kakinada Institute of Engineering & Technology (KIET)-II, Kakinada On 12th February, 2013, CSI student branch was inaugurated and seminar on Cloud Computing was organized. There was a talk delivered by Dr. C V S Murty, on research methodologies and also a talk by Mr. Sekhar Kammula, who spoke about social activity named I CARE I REACT. Mr. Naganand Rapaka spoke about cloud computing technologies.
REGION VI
G. H. Raisoni Institute of Information Technology, Nagpur On the occasion of inauguration of CSI student branch on 6th of February, 2013, motivational speech was given by Mr. Pratap Shukla, on Journey from Learning Software Programmer to Expert Solution Architect. Other topics covered in the inauguration seminar were - Programming Introductory Tools, Beginner Programmer Tools, Language Developer Tools, and Solution Architect Tools.
REGION VII
E.G.S. Pillay Engineering College, Nagore, Nagapattinam CSI Student Branch was launched at EGS Pillay Engineering College, on March 05, 2013. Participating in the event as the Chief Guest, Mr. Ramasamy, Regional Vice President, explained about the activities to be taken up by branch, and techniques that would help students to face examinations and interviews successfully.
REGION VII
Sethu Institute of Technology, Sivakasi CSI student branch inauguration was organized on February 27, 2013. Mr. Ramasamy, Vice Chairman, of CSI Chapter Chennai, gave inaugural address and highlighted opportunities for professional development.
CSI CS I Co Comm Communications mmun mm unic un ic cat atio ions ns s | Ap April A r l 2013 ri 201 13 | 50 0
www. ww w.cs w. csics i-in indi in dia di a.or a.or org g www.csi-india.org
Announcement
Programme Proposals
Pursuant to the above mission and objectives, CSI invites proposals from Chapters, Student Branches, Institutional members as well as individual members, for different kinds of international, national, regional, and state-level events for the year 2013-14, including, but not limited to: Technical Conferences (national/ international) Seminars Workshops Research Symposia Faculty Development Programmes Job Fairs Exhibitions lT initiation programmes for schools Quiz programmes Student conventions for college and school students Pilot programmes These events will be organized by Chapters, Student Branches or other organizational units of CSl, and supported by the entire CSI ecosystem, viz., CSI Headquarters CSI Educational Directorate RVPs & Divisional Chairs National, Regional and State-level Student Co-ordinators and other regional staff SIG Chairs (where appropriate) Chapters and Student Branches Partnering organizations (e.g., IEEE Computer Society, CDAC, PMI) Associated international/national
Proposal Guidelines
Kindly apply to vp@csi-india.org with the following particulars on or before 31 May 2013: Title of the event Type of event (e.g., Seminar/ Workshop/Conference) Hosting Unit(s) (e.g., Chapter/ Student Branch/Region/ Divisions/ SlGs) Duration Location Topics and outline of the event programmes Proposed benets to CSI and members Potential Partners and Sponsors Target audience and size Preliminary Budget (Revenue and Surplus) Kindly note that these events are expected to adhere to the provisions of the CSI Conferences Manual (please refer to the CSI Website) as applicable! CSI proposes to recognize the contribution of members and institutions for organizing events in the form of awards. Criteria and category of awards n will be announced soon.
Registered with Registrar of News Papers for India - RNI 31668/78 Regd. No. MH/MR/N/222/MBI/12-14 Posting Date: 10 & 11 every month. Posted at Patrika Channel Mumbai-I Date of Publication:10 & 11every month
If undelivered return to : Samruddhi Venture Park, Unit No.3, 4th oor, MIDC, Andheri (E). Mumbai-400 093
CSI 2013
48th Annual Convention of the Computer Society of India Hosted by: CSI Visakhapatnam Chapter In association with Visakhapatnam Steel Plant Theme: ICT and Critical Infrastructure Dates: 13-15, December 2013 Venue: Hotel Novotel, Visakhapatnam Call For Papers / Participation
Introduction: CSI-2013, the 48th Annual Convention of Computer Society of India (CSI), is being organized by CSI Visakhapatnam Chapter, in association with Rastriya Ispat Nigam Limited, Visakhapatnam Steel Plant to bring together researchers, engineers, developers, and practitioners from academia and industry working in all interdisciplinary areas of information system engineering and computing, Innovative IT professionals from government establishments to small, medium & big enterprises, from non-government organizations to multi-national companies to share the experience, exchange ideas and update their knowledge in latest developments in emerging areas. Following the big successes of previous conferences, CSI, Visakhapatnam Chapter is set to conduct the First Annual Convention at Visakhapatnam, CSI-2013 that will serve as a forum for discussions on the state-of-the-art research, development and implementations of ICT applications. The progress and the growth of any country depend on the Infrastructure and ICT becoming pervasive has a crucial role in managing the Infrastructure. Keeping this in mind, the theme for CSI-2013 has been selected as ICT and Critical Infrastructure. The deliberations will focus on this aspect and cover Innovative ways to deliver business values, optimize business processes and enable inclusive growth. It will also focus on proven IT governance, standards, practices, design & tools that lead to fast development and information ow to the user. Invitation: We invite authors to submit papers reecting original research work and practical experiences in the areas of interest to the convention. Invitation is extended to CEOs/CIOs, IT Professionals, IT Users, academicians, researchers, students, and members of the CSI to attend as delegates in this convention. Software rms, Industries and business houses are invited to participate in the convention and present and exhibit their products and services. CSI 2013, invites papers of original research and pertaining to ICT and Critical Infrastructure and on the following topics (but not limited to): * ICT use in Critical Infrastructure (CI) * Security Challenges in using ICT in CI * Wireless and Mobility technologies in the Control Loop * ICT in Steel Industry * ICT in Heavy & Manufacturing Industry * ICT in Process Industry* ICT in BFSI * ICT in Transportation * ICT in Education * ICT in Telecom * ICT in Healthcare * ICT in E-Commerce * ICT in Maritime - Navy, Ship Building, Ocean understanding * ICT in Rural Areas * ICT in eGovernance * Programming Paradigms for CI * Designing Applications for CI * ICT and Cyber Physical Systems in coastal areas * Role of OTT in CI * Synergistic Policy Framework for ICT and CI * Coexistence of OTT, Cloud, and Social Networks * Big Data Analysis and CI * Machine Intelligence * Soft Computing Applications * AI and Nano Computing * Geo informatics and Environment * Bio-informatics * Software Engineering * IT Security, Forensics and Cyber Crime We also invite proposals for workshops, pre-conference tutorials and doctoral consortium. Publication: Prospective authors are invited to submit paper(s) not exceeding 8 pages written in A4 size, and as per the AISC, Springer format on any one of the tracks listed above. The proceedings will be published by AISC series of Springer. Important Dates:
Submission of Full Manuscript Notication of Acceptance Camera ready copy Submission of Tutorial/Workshop Proposals Registration Starts 15 July 2013 15 Aug 2013 31 Aug 2013 30 July 2013 31 Aug 2013
Paper Submission for CSI - 2013 https:/ /www.easychair.org/conferences/?conf=csi2013 For more details please visit http:/ /www.csi-2013.org
Address for Communication Paramata Satyanarayana Convener, CSI-2013 Sr. Manager, Central Computer Center Visakhapatnam Steel Plant, Visakhapatnam 530 031 Mobile: +91 9949556989 Email: s_paramata@vizagsteel.com convener@csi-2013.org Organizing Committee Chair Sri TK Chand, D(C), RINL Programme Committee Chair Prof PS Avadhani, AUCE (A) Finance Committee Chair Sri GN Murthy, ED (F&A), VSP