carnegie mellon ai, sensing, and optimized information gathering: trends and directions carlos...

46
Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg, Brendan McMahan, Ajit Singh, and others…

Upload: kimberly-adelia-allen

Post on 24-Dec-2015

213 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Carnegie Mellon

AI, Sensing, and Optimized Information Gathering:

Trends and Directions

Carlos Guestrin

joint work with:

and:Anupam Gupta, Jon Kleinberg,

Brendan McMahan, Ajit Singh, and others…

Page 2: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Monitoring algal blooms

Algal blooms threaten freshwater4 million people without water1300 factories shut down$14.5 billion to clean upOther occurrences in Australia, Japan, Canada, Brazil, Mexico, Great Britain, Portugal, Germany …

Growth processes still unclear [Carmichael]Need to characterize growth in the lakes, not in

the lab!

Tai Lake China10/07 MSNBC

Page 3: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Can only make a limited number of measurements!

Dep

th

Location across lake

Monitoring rivers and lakesNeed to monitor large spatial phenomena

Temperature, nutrient distribution, fluorescence, …

Predict atunobserved

locations

NIMSKaiseret.al.

(UCLA)

Color indicates actual temperature Predicted temperature

Use robotic sensors tocover large areas

Where should we sense to get most accurate predictions?

[Singh, Krause, G., Kaiser ‘07]

Page 4: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Water distribution networks

Simulator from EPA

Water distribution in a city very complex systemPathogens in water can affect thousands (or millions) of peopleCurrently: Add chlorine to the source and hope for the best

Chlorine

ATTACK!could deliberately

introduce pathogen

Page 5: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Monitoring water networks[Krause, Leskovec, G., Faloutsos, VanBriesen ‘08]

Contamination of drinking watercould affect millions of people

Place sensors to detect contaminations“Battle of the Water Sensor Networks” competitionWhere should we place sensors

to detect contaminations quickly ?

Sensors

Simulator from EPA Hach Sensor~$14K

Page 6: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Sensing problemsWant to learn something about the state of the world

Detect outbreaks, predict algal blooms …

We can choose (partial) observations…Place sensors, make measurements, …

… but they are expensive / limitedhardware cost, power consumption, measurement time …

Want cost-effectively get most useful information!

Fundamental problem:What information should I use to learn

?

Page 7: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Related work

Sensing problems considered inExperimental design (Lindley ’56, Robbins ’52…), Spatial

statistics (Cressie ’91, …), Machine Learning (MacKay ’92, …), Robotics (Sim&Roy ’05, …), Sensor Networks (Zhao et al ’04, …), Operations Research (Nemhauser ’78, …)

Existing algorithms typicallyHeuristics: No guarantees! Can do arbitrarily badly.Find optimal solutions (Mixed integer programming, POMDPs):Very difficult to scale to bigger problems.

Page 8: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

This talk

Theoretical: Approximation algorithms that have theoretical guarantees and scale to large problems

Applied: Empirical studies with real deployments and large datasets

Page 9: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Model-based sensingModel predicts impact of contaminations

For water networks: Water flow simulator from EPAFor lake monitoring: Learn probabilistic models from data (later)

For each subset AÍV compute “sensing quality” F(A)

S2

S3

S4S1 S2

S3

S4

S1

High sensing quality F(A) = 0.9Low sensing quality F(A)=0.01

Model predictsHigh impact

Medium impactlocation

Lowimpactlocation

Sensor reducesimpact throughearly detection!

S1

Contamination

Set V of all network junctions

Page 10: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Robust sensing Complex constraints

Sequential sensing

Optimizing sensing / Outline

Sensing locationsSensing quality

Sensing budgetSensing cost

Sensor placement

Page 11: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Sensor placementGiven: finite set V of locations, sensing quality FWant: A*Í V such that

Typically NP-hard!

How well can this simple heuristic do?

S1

S2

S3

S4

S5

S6Greedy algorithm:

Start with A =Ø ;For i = 1 to k

s* := argmaxs F(A {s})A := A {s*}

Page 12: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

2 4 6 8 100.5

0.6

0.7

0.8

0.9

Number of sensors placed

Pop

ulat

ion

affe

cted

Performance of greedy algorithm

Greedy score empirically close to optimal. Why?

Small subset of Water networks

data

2 4 6 8 100.5

0.6

0.7

0.8

0.9

Number of sensors placed

Pop

ulat

ion

affe

cted

Greedy

Optimal

Pop

ula

tion

pro

tecte

d

(hig

her

is b

ett

er)

Number of sensors placed

Page 13: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

S2S3

S4S1

Key property: Diminishing returns

S2

S1

S’

Placement A = {S1, S2} Placement B = {S1, S2, S3, S4}

Adding S’ will help a lot!

Adding S’ doesn’t help muchNew

sensor S’

B . . . . . A S’

S’

+

+

Large improvement

Small improvement

For AÍB, F(A {S’}) – F(A) ≥ F(B {S’}) – F(B)

Submodularity:

Theorem [Krause, Leskovec, G., Faloutsos, VanBriesen ’08]:Sensing quality F(A) in water networks is

submodular!

Page 14: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

One reason submodularity is useful

Theorem[Nemhauser et al ‘78]

Greedy algorithm gives constant factor approximation

F(Agreedy)≥ (1-1/e) F(Aopt)

Greedy algorithm gives near-optimal solution!Guarantees best possible unless P = NP!

Many more reasons, sit back and relax…

~63%

Page 15: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

People sit a lotActivity recognition inassistive technologiesSeating pressure as user interface Equipped with

1 sensor per cm2!

Costs $16,000!

Can we get similar accuracy with fewer,

cheaper sensors?

Leanforward

SlouchLeanleft

82% accuracy on 10 postures! [Zhu et al]

Building a Sensing Chair [Mutlu, Krause, Forlizzi, G., Hodgins ‘07]

Page 16: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

How to place sensors on a chair?

Sensor readings at locations V as random variablesPredict posture Y using probabilistic model P(Y,V)Pick sensor locations A*ÍV to minimize entropy:

Possible locations V

Theorem: Information gainis submodular!*[UAI’05]

*See store for details

Accuracy CostBefore 82% $16,000 After 79% $100

Placed sensors, did a user study:

Similar accuracy at <1% of cost!

Page 17: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Battle of the Water Sensor Networks Competition

Real metropolitan area network (12,527 nodes)Water flow simulator provided by EPA3.6 million contamination eventsMultiple objectives: Detection time, affected population, …Place sensors that detect well “on average”

Page 18: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

BWSN Competition results

13 participantsPerformance measured in 30 different criteria

0

5

10

15

20

25

30

Tota

l S

core

Hig

her

is b

ett

er

Our

app

roac

h

Berry

et. a

l.

Dor

ini e

t. a

l.

Wu

& W

alsk

i

Ost

feld

& S

alom

ons

Prop

ato

& P

iller

Elia

des & P

olyc

arpo

u

Hua

ng e

t. a

l.

Gua

n et

. al.

Ghi

mire

& B

arkd

oll

Trac

htm

an

Gue

li

Prei

s & O

stfe

ld

E

E

D DG

GG

GG

H

H

H

G: Genetic algorithm

H: Other heuristic

D: Domain knowledge

E: “Exact” method (MIP)

24% better performance than runner-up!

Page 19: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Simulated all on 2 weeks / 40 processors152 GB data on disk Very accurate sensing quality

, 16 GB in main memory (compressed)

Using “lazy evaluations”:1 hour/20 sensorsDone after 2 days! Advantage through theory and

engineering!

Low

er

is b

ett

er 30 hours/20 sensors

6 weeks for all30 settings

3.6M contaminations

Very slow evaluation of F(A)

1 2 3 4 5 6 7 8 9 100

100

200

300

Number of sensors selected

Runn

ing

time

(min

utes

)

Exhaustive search(All subsets)

NaiveGreedy

ubmodularity to the rescue:

What was the trick?

Fast Greedy

Page 20: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Robustness against adversaries

Unified viewRobustness to change in parameters Robust experimental designRobustness to adversaries

SATURATE: A simple, but very effective algorithm for robust sensor placement

If sensor locations are known, attack vulnerable locations

[Krause, McMahan, G., Gupta ‘07]

Page 21: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

What about worst-case?

S2

S3

S4S1

Knowing the sensor locations, an adversary contaminates here!

Where should we place sensors to quickly detect in the worst case?

Very different average-case score,Same worst-case score

S2

S3

S4

S1

Placement detects well on “average-case”

(accidental) contamination

Page 22: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Optimizing for the worst case

Contamination at node s Sensors A

Fs(A) is high

Contamination at node r

Fr(A) is low

Fs(B) is low

Fr(B) is high

Sensors B

Fr(C) is high

Fs(C) is high

Sensors C

Separate utility function Fi with each contamination i

Fi(A) = impact reduction by sensors A for contamination i

Want to solve

Each of the Fi is submodular

Unfortunately, mini Fi

not submodular!How can we solve this robust sensing problem?

Page 23: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

How does the greedy algorithm do?

Theorem [NIPS ’07]: The problem max|A|≤ k mini Fi(A) does not admit any approximation unless P=NP

Optimalsolution

Greedy picks first

Then, canchoose only

or

Greedy does arbitrarily badly. Is there something better?

V={ , , }

Can only buy k=2

Greedy score:Optimal score: 1

Set A F1 F2 mini Fi

1 0 00 2 0

1 2

1 2 1

Hence we can’t find any approximation algorithm.

Or can we?

Page 24: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Alternative formulation

If somebody told us the optimal value,

can we recover the optimal solution A*?

Need to find

Is this any easier?

Yes, if we relax the constraint |A| ≤ k

Page 25: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Solving the alternative problem

Trick: For each Fi and c, define truncation

c

|A|

Fi(A)

F’i,c(A)

Same optimal solutions!Solving one solves the other

Non-submodular Don’t know how to solve

Submodular!Can use greedy!

Problem 1 (last slide) Problem 2

Page 26: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Back to our example

Guess c=1First pick Then pick

Optimal solution!

How do we find c?Do binary search!

Set A F1 F2 mini Fi F’avg,1

1 0 0 ½0 2 0 ½ 1 (1+)/2

2 (1+)/2

1 2 1 1

Page 27: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Saturate Algorithm [NIPS ‘07]Given: set V, integer k and submodular functions F1,…,Fm

Initialize cmin=0, cmax = mini Fi(V)

Do binary search: c = (cmin+cmax)/2

Greedily find AG such that F’avg,c(AG) = c

If |AG| ≤ k: increase cmin

If |AG| > k: decrease cmax

until convergence

Truncationthreshold(color)

Page 28: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Theoretical guarantees

Theorem:If there were polytime algorithm with better factor <, then NP DTIME(nlog log n)

Theorem: Saturate finds a solution AS such that

mini Fi(AS) ≥ OPTk and |AS| ≤ k

where OPTk = max|A|≤k mini Fi(A)

= 1 + log maxsi Fi({s})

Theorem: The problem max|A|≤ k mini Fi(A) does not admit any approximation unless P=NP

Page 29: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Example: Lake monitoring

Monitor pH values using robotic sensor

Position s along transect

pH

valu

e

True (hidden) pH values

Prediction at unobservedlocations

transect

Where should we sense to minimize our maximum error?

Use probabilistic model(Gaussian processes)

to estimate prediction error

(often) submodular[Das & Kempe ’08]

Var(s | A)

Robust sensing problem!

Observations A

Page 30: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Comparison with state of the art

Algorithm used in geostatistics: Simulated Annealing[Sacks & Schiller ’88, van Groeningen & Stein ’98, Wiens ’05,

…]7 parameters that need to be fine-tuned

Environmental monitoring

bett

er

0 20 40 600

0.05

0.1

0.15

0.2

0.25

Number of sensors

Max

imum

mar

gina

l var

ianc

e

Greedy

0 20 40 600

0.05

0.1

0.15

0.2

0.25

Number of sensors

Max

imum

mar

gina

l var

ianc

e

Greedy

SimulatedAnnealing

0 20 40 600

0.05

0.1

0.15

0.2

0.25

Number of sensors

Max

imum

mar

gina

l var

ianc

e

Greedy

SimulatedAnnealing

Saturate

Precipitation data

0 20 40 60 80 1000.5

1

1.5

2

2.5

Number of sensors

Max

imu

m m

arg

inal

var

ian

ce

Greedy

Saturate

SimulatedAnnealing

Saturate is competitive & 10x fasterNo parameters to tune!

Page 31: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Saturate

Results on water networks

60% lower worst-case detection time!

Water networks

500

1000

1500

2000

2500

3000

Number of sensors

Maxim

um

dete

cti

on

tim

e (

min

ute

s)

Low

er

is b

ett

er

No decreaseuntil allcontaminationsdetected!

0 10 200

Greedy

SimulatedAnnealing

Page 32: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Summary so far

Submodularity in sensing optimization

Greedy is near-optimal

Robust sensingGreedy fails badly

Saturate is near-optimal

Path planningCommunication constraints

Constrained submodular optimizationpSPIEL gives strong guarantees

Sequential sensingExploration Exploitation Analysis

All these applications involve physical sensing

Now for something completely different

Let’s jump from water…

Page 33: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

… to the Web!

You have 10 minutes each day for reading blogs / news.

Which of the million blogs should you read?

Page 34: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Tim

e

Information cascade

Which blogs should we read to learn about big cascades early?

Learn aboutstory after us!

Information Cascades[Leskovec, Krause, G., Faloutsos, VanBriesen ‘07]

Page 35: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Water vs. Web

In both problems we are givenGraph with nodes (junctions / blogs) and edges (pipes / links)Cascades spreading dynamically over the graph (contamination / citations)

Want to pick nodes to detect big cascades early

Placing sensors inwater networks

Selectinginformative blogsvs.

In both applications, utility functions submodular

Page 36: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Performance on Blog selection

Outperforms state-of-the-art heuristics700x speedup using submodularity!

Blog selection

Low

er

is b

ett

er

1 2 3 4 5 6 7 8 9 10

0

100

200

300

400

Number of blogs selected

Ru

nn

ing

tim

e (

secon

ds)

Exhaustive search(All subsets)

Naivegreedy

Fast greedy

Blog selection~45k blogs

Hig

her

is b

ett

er

Number of blogs

Cascad

es c

ap

ture

d

0 20 40 60 80 1000

0.1

0.2

0.3

0.4

0.5

0.6

0.7Greedy

In-linksAll outlinks

# Posts

Random

Page 37: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Naïve approach: Just pick 10 best blogsSelects big, well known blogs (Instapundit, etc.)These contain many posts, take long to read!

Taking “attention” into account

Casc

ades

cap

ture

d

Number of posts (time) allowed

x 104

Cost/benefitanalysis

Ignoring cost

Cost-benefit optimization picks summarizer blogs!

Page 38: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Predicting the “hot” blogs

Jan Feb Mar Apr May0

200

#detectio

ns

Greedy

Jan Feb Mar Apr May0

200

#detectio

ns

Saturate

Detects on training set

Greedy on historicTest on future

Poor generalization!Why’s that?

0 1000 2000 3000 40000

0.05

0.1

0.15

0.2

0.25 Greedy on futureTest on future

“Cheating”

Cascad

es c

ap

ture

d

Number of posts (time) allowed

Detect wellhere!

Detect poorlyhere!

Want blogs that will be informative in the futureSplit data set; train on historic, test on future

Blog selection “overfits”to training data!

Let’s see whatgoes wrong here.

Want blogs thatcontinue to do well!

Page 39: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Robust optimization

Jan Feb Mar Apr May0

200

#detectio

ns

Greedy

Jan Feb Mar Apr May0

200

#detectio

ns

Saturate

Jan Feb Mar Apr May0

200

#detectio

ns

Greedy

Jan Feb Mar Apr May0

200

#detectio

ns

Saturate

Detections using Saturate

F1(A)=.5

F2 (A)=.8

F3 (A)=.6

F4(A)=.01

F5 (A)=.02

Optimizeworst-case

Fi(A) = detections in interval i

“Overfit” blog selection A

“Robust” blog selection A*

Robust optimization Regularization!

Page 40: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Predicting the “hot” blogs

Greedy on historicTest on future

Robust solutionTest on future

0 1000 2000 3000 40000

0.05

0.1

0.15

0.2

0.25 Greedy on futureTest on future

“Cheating”S

en

sin

g q

uality

Number of posts (time) allowed

50% better generalization!

Page 41: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Summary

Submodularity in sensing optimization

Greedy is near-optimal

Robust sensingGreedy fails badly

Saturate is near-optimal

Path planningCommunication constraints

Constrained submodular optimizationpSPIEL gives strong guarantees

Sequential sensingExploration Exploitation Analysis

Constrained optimization better use of

“attention”

Robust optimization better generalization

Page 42: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

AI-complete dreamRobot that saves the world

Robot that cleans your room

But…It’s definitely useful, but…

Really narrow

Hardware is a real issueWill take a while

What’s an “AI-complete” problem that will be useful to a huge number of people in the next 5-10 years?What’s a problem accessible to a large part of AI community?

Page 43: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

What makes a good AI-complete problem?

A complete AI-system:Sensing: gathering information from the worldReasoning: making high-level conclusions from informationActing: making decisions that affect the dynamics of the world and/or the interaction with the user

But alsoHugely complexCan get access to real dataCan scale up and layer upCan make progressVery cool and exciting

Data gathering can lead to good, accessible and cool AI-complete

problems

Page 44: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Factcheck.orgTake a statement

Collect information from multiple sources

Evaluate quality of sources

Connect them

Make a conclusion AND provide an analysis

Page 45: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

Automated fact checkingQuery

Fact or Fiction

?Conclusion

and Justification

Active user feedback on sources and proof

Web

Models

InferenceCan lead to very cool “AI-

complete” problem, useful, and can make progress in short

term!

Page 46: Carnegie Mellon AI, Sensing, and Optimized Information Gathering: Trends and Directions Carlos Guestrin joint work with: and: Anupam Gupta, Jon Kleinberg,

ConclusionsSensing and information acquisition problems are important and ubiquitousCan exploit structure to find provably good solutionsObtain algorithms with strong guaranteesPerform well on real world problems

Could help focus on a cool “AI-complete”

problem