1、Artificial Intelligence and the Internet,Edward Brent University of Missouri Columbia and Idea Works, Inc. Theodore Carnahan Idea Works, Inc.,Overview,Objective Consider how AI can be (and in many cases is being) used to enhance and transform social research on the Internet Framework intersection of
2、 AI and research issues View Internet as a source of data whose size and rate of growth make it important to automate much of the analysis of data,Overview (continued),We discuss a leading AI-based approach, the semantic web, and an alternative paradigmatic approach, and the strengths and weaknesses
3、 of each We explore how other AI strategies can be used including intelligent agents, multi-agent systems, expert systems, semantic networks, natural language understanding, genetic algorithms, neural networks, machine learning, and data mining We conclude by considering implications for future rese
4、arch,Key Features of the Internet,Decentralized Few or no standards for much of the substantive content Incredibly diverse information Massive and growing rapidly Unstructured data,The Good News About the Internet,A massive flow of data Digitized A researchers dream,The Bad News,A massive flow of da
5、ta Digitized A researchers nightmare,Data Flows,The Internet provides many examples of data flows. A data flow is an ongoing flux of new information, often from multiple sources, and typically large in volume. Data flows are the result of ongoing social processes in which information is gathered and
6、/or disseminated by humans for the assessment or consumption by others. Not all data flows are digital, but all flows on the Internet are. Data flows are increasingly available over the internet. Examples of data flows include News articles Published research articles eMail Medical records Personnel
7、 records Articles submitted for publication Research proposals Arrest records Birth and death records,Data Flows vs Data Sets,Data flows are fundamentally different from the data sets with which most social scientists have traditionally worked.,The Need for Automating Analysis,Together, the tremendo
8、us volume and rate of growth of the Internet, and the prevalence of ongoing data flows make automating analysis both more important and more cost-effective. Greater cost savings result from automated analysis with very large data sets Ongoing data flows require continuing analysis and that also make
9、s automation cost-effective,AI and Automating Research,Artificial Intelligence strategies offer a number of ways to automate research on the Internet. We,Contemporary Social Research on the Web,Formulate the research problem Search for and sample web sites containing relevant data Process, format, s
10、tore data for analysis Develop a coding scheme Code web pages for analysis Conduct analyses,Strengths and Weaknesses of Contemporary Approach,May use qualitative or quantitative programs to assist with the coding and analysis Advantages Versatile Gives researcher much control Disadvantages Coding sc
11、hemes often not shared, requiring more effort, making research less cumulative and less objective Expensive and time-consuming Unlikely to keep up with rapidly changing data in data flows Not cost-effective for ongoing analysis and monitoring,The Semantic Web,The semantic web is an effort to build i
12、nto the World Wide Web tags or markers for data along with representations of the semantic meaning of those tags (Berners-Lee and Lassila, 2001; Shadbolt, Hall and Berners-Lee, 2006). The semantic web will make it possible for computer programs to recognize information of a specific type in any of m
13、any different locations on the web and to “understand” the semantic meaning of that information well enough to reason about it. This will produce interoperability the ability of different applications and databases to exchange information and to be able to use that information effectively across app
14、lications. Such a web can provide an infrastructure to facilitate and enhance many things including social science research.,Implementing the Semantic Web,The Semantic Web: What Can It Do?,Illustrate briefly,AI Strategies and the Semantic Web,Several components of the semantic web make use of artifi
15、cial intelligence (AI) strategies,Strengths of the Semantic Web,Fast and efficient to develop Most coding done by web developers one time and used by everyone Fast and efficient to use Intelligent agents can do most of the work with little human intervention Structure provided makes it easier for co
16、mputers to process Can take advantage of distributed processing and grid computing Interoperability Many different applications can access and use information from throughout the web,Weaknesses of the Semantic Web (Pragmatic Concerns),Seeks to impose standardization on a highly decentralized process
17、 of web development Requires cooperation of many if not all developers Imposes the double burden of expressing knowledge for humans and for computers How will tens of millions of legacy web sites be retrofitted? What alternative procedures will be needed for noncompliant web sites? Major forms of da
18、ta on the web are provided by untrained users unlikely to be able to markup for the semantic web E.g., blogs, input to online surveys, emails,Weaknesses of the Semantic Web (Fundamental Concerns),Assumes there is a single ontology that can be used for all web pages and all users (at least in some do
19、main). For example, a standard way to markup products and prices in commercial web sites could make it possible for intelligent agents to search the Internet for the best price for a particular make and model of car. This assumption may be inherently flawed for social research for two reasons. 1) Mu
20、ltiple paradigms - What ontology could code web pages from multiple competing paradigms or world views (Kuhn, 1969). If reality is socially constructed, and “beauty is in the eye of the beholder” how can a single ontology represent such diverse views? 2) Competing interests What if developers of web
21、 pages have political or economic interests at odds with some of the viewers of those web pages?,Multiple Perspectives,Chomskys deep structure vs subtexts,Contested terms,Paradigmatic Approach,We describe an alternative approach to the semantic web, one that we believe may be more suitable for many
22、social science research applications. Recognizes there may be multiple incompatible views of data Data structure must be imposed on data dynamically by the researcher as part of the research process (in contrast to the semantic web which seeks to build an infrastructure of web pages with data struct
23、ure pre-coded by web developers),Paradigmatic Approach (continued),Relies heavily on natural language processing (NLP) strategies to code data. NLP capabilities are not already developed for many of these research areas and must be developed. Those NLP procedures are often developed and refined usin
24、g machine learning strategies.We will compare the paradigmatic approach to traditional research strategies and the Semantic Web for important research tasks.,Example Areas Illustrating the Paradigmatic Approach,Event analysis in international relations Essay grading Tracking news reports on social i
25、ssues or for clients E.g., Campaigns, Corporations, Press agentsEach of these areas illustrate significant data flows. These areas and programs within them illustrate elements of the paradigmatic approach. Most do not yet employ all the strategies.,Essay Grading,These are programs that allow student
26、s to submit essays using the computer then a computer program examines the essays and computes a score for the student. Some of the programs also provide feedback to the student to help them improve. These programs are becoming more common for standardized assessment tests and classroom applications
27、. Examples of programs SAGrader E-rater C-rater Intelligent Essay Assessor Criterion These programs illustrate large ongoing data flows and generally reflect the paradigmatic approach.,Digitizing Data,The first step in any computer analysis must be converting relevant data to digital form where it i
28、s expressed as a stream of digits that can be transmitted and manipulated by computers These two approaches both rely on web page developers to digitize information. This gives them a distinct advantage over traditional research where digitizing data can be a major hurdle.,Essay Grading: Digitizing
29、Data,Digitizing Papers replaced with digital submissions SAGrader, for example, has students submit their papers over the Internet using standard web browsers. Digitizing often still a major hurdle limiting use Access issues Security concerns,Data Conversions,Essay Grading: Converting Data,Data conv
30、ersion Where essays are submitted on paper, optical character recognition (OCR) or handwriting recognition programs must be used to convert to digitized text. Standardized testing programs often face this issue,Encoding Data,Essay Grading: Coding,Essay grading programs employ a wide array of strateg
31、ies for recognizing important features in essays. Intelligent Essay Assessor (IEA) employs a purely statistical approach, latent semantic analysis (LSA). This approach treats essays like a “bag of words” using a matrix of word frequencies by essays and factor analysis to find an underlying semantic
32、space. It then locates each essay in that space and assesses how closely it matches essays with known scores. E-rater uses a combination of statistical and linguistic approaches. It uses syntactic, discourse structure, and content features to predict scores for essays after the program has been trai
33、ned to match human coders. SAGrader uses a strategy that blends linguistic, statistical, and AI approaches. It uses fuzzy logic to detect key concepts in student papers and a semantic network to represent the semantic information that should be present in good essays. All of these programs require l
34、earning before they can be used to grade essays in a specific domain.,Knowledge,Essay Grading: Knowledge,Most essay grading programs have very little in the way of a representation of theory or knowledge. This is probably because they are often designed specifically for grading essays and are not me
35、ant to be used for other purposes requiring theory, such as social science research. For example, C-rater, a program that emphasizes semantic content in essays, yet has no representation of semantic content other than as desirable features for the essay. The exception is SAGrader. SAGrader employs t
36、echnologies developed in a qualitative analysis program, Qualrus. Hence, SAGrader uses a semantic network to explicitly represent and reason about the knowledge or theory.,Analysis,The semantic web and paradigmatic approaches can take similar approaches to analysis.,Essay Grading: Analysis,All progr
37、ams produce scores, though the precision and complexity of the scores varies. Some produce explanations Most of these essay grading programs simply perform a one-time analysis (grading) of papers. However some of them, such as SAGrader, provide for ongoing monitoring of student performance as studen
38、ts revise and resubmit their papers. Since essays presented to the programs are already converted into standard formats and are submitted to a central site for processing, there is no need for the search and retrieval capabilities of intelligent agents,Advantages of Paradigmatic Approach,Suitable fo
39、r multiple-paradigm fields Suitable for contested issues Does not require as much infrastructure development on the web Can be used for new views requiring different codes with little lag time,Disadvantages of Paradigmatic Approach,Relies heavily on NLP technologies that are still evolving May not b
40、e feasible in some or all circumstances Requires extensive machine learning Often requires additional data conversion for automated analysis Requires individual web pages to be coded once for each paradigm rather than a single time, hence increasing costs. (However, by automating this, costs are mad
41、e manageable) Current NLP capabilities are limited to problems of restricted scope. Instead of general-purpose NLP programs, they are better characterized as special-purpose NLP programs.,Structured Data,Structured data data stored in a computer in a manner that makes it efficient to examine A good
42、data structure does much of the work, making the algorithms required for some kinds of reasoning straightforward, even trivial. Examples of structured data include data stored in spreadsheets, statistical programs, and data bases. Unstructured data data stored in a manner that does not make it effic
43、ient to examine Examples of unstructured data include newspaper articles, blogs, interview transcripts, and graphics files. A structured unstructured dichotomy is an oversimplification Data well-structured for some purposes may not be well-structured for other purposes. For viewing by humans E.g., p
44、hotographs, protected pdf files For processing by programs E.g., text, doc, html Marked for analysis (semantic web),Event Analysis,How is this data flow?,Event Analysis,Schrodts discusison of various coding schemes,Discussion and Conclusions,Both semantic web and paradigmatic approaches have advanta
45、ges and disadvantages Codes on semantic web could facilitate coding by paradigmatic-approach programs Where there is much consensus the single coding for the semantic web could be sufficient While the infrastructure for the semantic web is still in development the paradigmatic approach could facilitate analysis of legacy data The paradigmatic approach could be used to build out the infrastructure for the semantic web,