




版權說明:本文檔由用戶提供并上傳,收益歸屬內容提供方,若內容存在侵權,請進行舉報或認領
文檔簡介
Explainability&
CommonRobustness姜育剛,馬興軍,吳祖煊
1.
What
is
Machine
Learning
2.
Machine
Learning
Paradigms3.
Loss
FunctionsRecap:
week
14.
Optimization
MethodsMachine
Learning
Pipelinesetuptheinputsetuptheoptimisersetupthelossregularizationmakesdecisionregionsmootherlandscape
ofalossfunction,itvariesw.r.t.data,thefunctionitselfMachine
Learning
Pipelinesetuptheinputsetuptheoptimisersetupthelossregularizationmakesdecisionregionsmootherlandscape
ofalossfunction,itvariesw.r.t.data,thefunctionitselfModel?Deep
Neural
Networks/neural-network-zoo/;/articles/cc-machine-learning-deep-learning-architectures/Feed-Forward
Neural
NetworksFeed-ForwardNeuralNetworks
(FNN)Fully
Connected
Neural
Networks
(FCN)Multilayer
Perceptron
(MLP)The
simplest
neural
networkFully-connectedbetweenlayersFordatathathasNOtemporalorspatialorder/ConvolutionalNeuralNetworksForimagesordatawithspatialorderCan
stack
up
to
>100
layers/Neurons
in
3
dimensionsNeurons
in
one
flat
layerRecurrent
Neural
Networks/~shervine/teaching/cs-230/cheatsheet-recurrent-neural-networksTraditional
RNNTransformersVaswani,Ashish,etal."Attentionisallyouneed."
Advancesinneuralinformationprocessingsystems
30(2017)Transformer:
a
new
type
of
DNNs
based
on
attentionEncoderDecoderSelf-Attention
Explained/illustrated-self-attention-2d627e33b20aCNN
ExplainedLearns
different
levels
of
representations/A
brief
history
of
CNNs:LeNet,1990sAlexNet,2012ZFNet,2013GoogLeNet,2014VGGNet,2014ResNet,2015InceptionV4,2016ResNeXt,
2017ViT,
2021AnImageisWorth16x16Words:TransformersforImageRecognitionatScale,
ICLR
2021Explainable
AI深度學習可解釋性學習機理推理機理泛化機理認知機理魯棒性學習過程學習結果決策依據推理機制泛化原因泛化條件認知科學認知啟發的智能普通魯棒性對抗魯棒性我們想要弄清楚下列問題:DNN是怎么學習的、學到了什么、靠什么泛化、在什么情況下行又在什么情況下不行?深度學習是否是真正的智能,與人類智能比誰更高級,它的未來是什么?是否存在大一統的理論,不但能解釋而且能提高?Methodological
PrinciplesVisualizationAblationContrastModelComponentLayerOperationNeuronSuperclassClassTraining/Test
setSubsetSampleTrainingInferenceTransferReverseHow
to
Understand
Machine
LearningLearning
is
the
process
of
empirical
risk
minimization
(ERM)Learning
MechanismTraining/Test
Error/AccuracyPrediction
Confidence
Explanation
via
observation:
just
plot!Wang
et
al.
SymmetricCrossEntropyforRobustLearningwithNoisyLabels,
ICCV
2019.Learning
MechanismParameter
dynamicsGradient
dynamicsExplanation
via
dynamics
and
informationTRADI:Trackingdeepneuralnetworkweightdistributions,
ECCV
2020;
Shwartz-ZivR,TishbyN.Openingtheblackboxofdeepneuralnetworksviainformation[J].arXiv:1703.00810,2017.Learning
MechanismDecision
boundary,
learning
process
visualizationExplanation
via
dynamics
and
informationhttps://distill.pub/2020/grand-tour/(March16,2020);
/Learning
MechanismData
influence/valuation:
how
a
training
sample
impacts
the
learning
outcome?UnderstandingBlack-boxPredictionsviaInfluenceFunctions,
ICML,
2018;
PruthiG,LiuF,KaleS,etal.Estimatingtrainingdatainfluencebytracinggradientdescent.NeurIPS,2020.Datashapley:Equitablevaluationof
data
formachinelearning,
ICML,
2019.Influence
FunctionData
ShapleyInfluence
FunctionHow
model
parameter
would
change
if
a
sample
z
is
removed
from
the
training
set?UnderstandingBlack-boxPredictionsviaInfluenceFunctions,
ICML,
2018;
目標:
Cook,R.D.andWeisberg,S.Residualsandinfluenceinregression.NewYork:ChapmanandHall,1982
所以:
Training
Data
InfluenceHow
model
loss
on
z’
would
change
if
update
on
a
sample
z?PruthiG,LiuF,KaleS,etal.Estimatingtrainingdatainfluencebytracinggradientdescent.NeurIPS,2020First-order
approximation
of
the
above
(assuming
one
step
update
is
small)?Checkpoints
store
the
interim
updates所以:Understanding
the
Learned
ModelLoss
LandscapeDeep
featurest-SNE
plotMaaten
et
al.Visualizingdatausingt-SNE.
JMLR,
2008.https://distill.pub/2016/misread-tsne/?_ga=2.135835192.888864733.1531353600-1779571267.1531353600Understanding
the
Learned
ModelClass-wise
PatternsIntermediate
Layer
Activation
MapActivation/Attention
MapLi
et
al.
NeuralAttentionDistillation:ErasingBackdoorTriggersfromDeepNeuralNetwork,
ICLR
2021;
Zhao
etal.Whatdodeepnetslearn?class-wisepatternsrevealedintheinputspace.arXiv:2101.06898
(2021).One
predictive
pattern
for
each
classWhat
do
deep
nets
learn?Zhao,Shihao,etal."Whatdodeepnetslearn?class-wisepatternsrevealedintheinputspace."
arXiv:2101.06898
(2021).Goal:
understanding
knowledge
learned
by
a
model
of
a
particular
class.Method:
Extract
one
single
pattern
for
one
class,
then
what
this
pattern
would
be?
Other
considerations:
we
need
to
do
this
in
pixel
space,
as
they
are
more
interpretableHow
to
Find
the
Class-wise
Pattern:
a
canvas
imagePatterns
extracted
on
different
canvases
(red
rectangles)Class-wise
Patterns
RevealedPatterns
extracted
on
original,
non-robust,
robust
CIFAR-10and
patterns
of
adversarially
trained
modelsPredictive
power
of
different
sizes
of
patternsInference
MechanismClass
Activation
Map
(Grad-CAM)Guided
BackpropagationSelvaraju
etal.Grad-cam:Visualexplanationsfromdeepnetworksviagradient-basedlocalization.
ICCV
2017.Springenberg
et
al.
StrivingforSimplicity:TheAllConvolutionalNet,
ICLR
2015.Guided
BackpropagationSpringenbergetal.StrivingforSimplicity:TheAllConvolutionalNet,ICLR2015.
/@chinesh4/generalized-way-of-interpreting-cnns-a7d1b0178709ReLU
forward
passReLU
backward
passDeconvolution
for
ReLUGuided
BackpropagationClass
Activation
Mapping
(CAM)Zhou
et
al.LearningDeepFeaturesforDiscriminativeLocalization.CVPR,2016.
/@chinesh4/generalized-way-of-interpreting-cnns-a7d1b0178709GAP:
Global
Average
PoolingGrad-CAMB.Zhou,A.Khosla,L.A.,A.Oliva,andA.Torralba.LearningDeepFeaturesforDiscriminativeLocalization.InCVPR,2016;
/@chinesh4/generalized-way-of-interpreting-cnns-a7d1b0178709Grad-CAM
is
a
generalization
of
CAMCompute
neuron
importance:
Weighted
combination
of
activation
map,
then
interpolation:LIMELocalInterpretableModel-agnosticExplanations(LIME)Ribeiro
et
al.“Whyshoulditrustyou?”Explainingthepredictionsofanyclassifier.“
SIGKDD,
2016./marcotcr/lime
Integrated
GradientsSundararajanM,TalyA,YanQ.Axiomaticattributionfordeepnetworks,
ICML,2017./TianhongDai/integrated-gradient-pytorch
Integrate
the
gradients
along
the
wayCognitive
DistillationHuang
et
al.
DistillingCognitiveBackdoorPatternswithinanImage,
ICLR
2023MaskextractbycognitivedistillationUsefulandnon-usefulfeaturesUsefulfeatures:highlycorrelatedwiththetruelabelinexpectation,
soIfremoved,predictionchangeBackdoortriggerisausefulfeatureNon-usefulfeatures:notcorrelated
with
predictionIfremoved,predictiondoesnotchangeIlyas,Andrew,etal."Adversarialexamplesarenotbugs,theyarefeatures.”NeurIPS2019CognitiveDistillationObjective:distilltheminimalessenceofusefulfeaturesModelTotalVariationLossRandomnoisevectorOriginalimageMaskCognitivePatternCognitiveDistillationDistilledpatternsonbackdoored
samplesxcpmxHow
to
VerifyCognitivePatterns
are
EssentialBackdooredimageBinarizedmask{0,1}OriginalimageConstruct
simplified
backdoor
patterns:Backdoor
Patterns
Can
Be
Made
Simplerxcpmxxbd’Backdoor
Patterns
Can
Be
Made
SimplerSimplified
backdoor
patterns
also
work!L1Norm
Distributionofthe
Distilled
MaskDetect
Backdoor
SamplesAttacks:12backdoorattacksModels:ResNet-18,Pre-ActivationResNet-101,MobileNetv2,VGG-16,Inception,EfficientNet-b0Datasets:CIFAR-10/GTSRB/ImageNetsubsetEvaluation
metric:areaundertheROCcurve(AUROC)Detectionbaselines:Anti-BackdoorLearning(ABL)[2]ActivationClustering(AC)[3]Frequency[4]STRIP[5]SpectralSignatures[6]CD-L(logitslayer)andCD-F(lastactivationlayer)Superb
Detection
PerformanceCelebA
dataset:40binaryfacialattributes(gender,bald,andhaircolor)KnownbiasbetweengenderandblondhairApply
CDinthesamewayasbackdoordetectionSelectsubsetofsampleswithlowL1normExamineattributesofthesubsetCalculatedistributionshiftbetweensubsetandthefulldatasetDiscover
Biases
in
Facial
Recognition
ModelsDiscover
Biases
in
Facial
Recognition
ModelsMasks
distilled
for
predicting
each
attributeDiscover
Biases
in
Facial
Recognition
ModelsGeneralization
MechanismConvergenceGeneralizationDeep
Learning
TheoryConvergenceConvex
(Linear
model)Nonconvex
(DNN)Saddle
pointGeneralizationTraining
time‘Cat’Test
time‘Cat’?Traditional
theory:
simpler
model
is
better,
more
data
is
betterGeneralization
Theory/~ninamf/ML11/lect1117.pdf;/watch?v=zlqQ7VRba2YComponents
of
Generalization
Error
Boundsgeneralizationerror
empiricalerror
hypothesisclasscomplexity
confidencesample
sizeRHS:
for
all
terms,
the
lower
the
better:
small
training
errorsimpler
model
classmore
samples
less
confidenceGeneralization
TheoryZhang
et
al.
Understandingdeeplearningrequiresrethinkinggeneralization.
ICLR
2017.Small
training
error≠low
generalization
errorZero
training
error
was
achieved
on
purely
random
labels
(meaningless
learning)0
training
error
vs.
0.9
test
errorList
of
Existing
TheoriesRademacher
Complexity
bounds
(Bartlett
et
al.
2017)PAC-Bayes
bounds
(Dziugaite
and
Roy
2017)Information
bottleneck
(Tishby
and
Zaslavsky
2015)Neural
tangent
kernel/Lazy
training
(Jacot
et
al.
2018)Mean-field
analysis
(Chizat
and
Bach
2018)Doule
Descent
(Belkin
et
al.
2019)Entropy
SGD
(Chaudhari
et
al.
2019)/watch?v=zlqQ7VRba2YA
few
interesting
questions:Should
we
consider
the
role
of
data
in
generalization
analysis?Should
representation
quality
appear
in
the
generalization
bound?Generalization
is
about
math
(the
function
of
the
model)
or
knowledge?How
to
visualize
generalization?
Existing
approachestest
errorVisualization:
loss
landscape,
prediction
attribution,
etc.Training
->
test:
distribution
shift,
out-of-distribution
analysisNoisy
labels
in
test
data
–
questioning
data
quality
and
reliable
evaluationThe
remaining
questions:
how
generalization
happens?Math≠KnowledgeComputation
=
finding
patterns
or
understanding
the
underlying
knowledgeWhat
is
the
relation
of
computational
generalization
to
human
behavior?Cognitive
MechanismOpenAI
reveals
the
multimodal
neurons
in
CLIP/blog/multimodal-neurons/;/blog/clip/Cognitive
MechanismRitter
et
al.
CognitivePsychologyforDeepNeuralNetworks:AShapeBiasCaseStudy,
ICML,
2017cognitivepsychology
inspired
evaluation
of
DNNsshape
match
=
prob
means
shape
biasCognitive
MechanismGeirhos,Robert,etal."Shortcutlearningindeepneuralnetworks."
NatureMachineIntelligence
2.11(2020):665-673.DeepneuralnetworkssolveproblemsbytakingshortcutsCognitive
MechanismRajalingham,Rishi,etal.“Large-scale,high-resolutioncomparisonofthecorevisualobjectrecognitionbehaviorofhumans,monkeys,andstate-of-the-artdeepartificialneuralnetworks.”
JournalofNeuroscience
38.33(2018):7255-7269.
Rajalingham,Rishi,KailynSchmidt,andJamesJ.DiCarlo."Comparisonofobjectrecognitionbehaviorinhumanandmonkey."
JournalofNeuroscience
35.35(2015):12127-121
溫馨提示
- 1. 本站所有資源如無特殊說明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請下載最新的WinRAR軟件解壓。
- 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請聯系上傳者。文件的所有權益歸上傳用戶所有。
- 3. 本站RAR壓縮包中若帶圖紙,網頁內容里面會有圖紙預覽,若沒有圖紙預覽就沒有圖紙。
- 4. 未經權益所有人同意不得將文件中的內容挪作商業或盈利用途。
- 5. 人人文庫網僅提供信息存儲空間,僅對用戶上傳內容的表現方式做保護處理,對用戶上傳分享的文檔內容本身不做任何修改或編輯,并不能對任何下載內容負責。
- 6. 下載文件中如有侵權或不適當內容,請與我們聯系,我們立即糾正。
- 7. 本站不保證下載資源的準確性、安全性和完整性, 同時也不承擔用戶因使用這些下載資源對自己和他人造成任何形式的傷害或損失。
最新文檔
- 榆林能源科技職業學院《鋼琴基礎二》2023-2024學年第一學期期末試卷
- 合肥職業技術學院《幼兒園語言教育活動設計與指導》2023-2024學年第二學期期末試卷
- 皖西學院《康復溝通與交流2》2023-2024學年第二學期期末試卷
- 天津理工大學《看花識草認中藥》2023-2024學年第二學期期末試卷
- 嘉興南洋職業技術學院《藥品質量控制》2023-2024學年第二學期期末試卷
- 邵陽學院《新媒體平臺運營實戰企業》2023-2024學年第二學期期末試卷
- 贛南科技學院《藝術批評學》2023-2024學年第二學期期末試卷
- 廣東工業大學《學院通選課傳統文化藝術》2023-2024學年第一學期期末試卷
- 3C認證基礎知識課件
- 人教PEP版英語五年級下冊教學課件Unit 4 Part A 第二課時
- 內能的利用教學設計(復習) 人教版
- 智聯招聘行測筆試題庫
- 2024中考化學試題研究專題《實驗室廢液成分的探究及處理》 課件
- 三年級數學兩位數乘兩位數筆算題綜合考核訓練題大全附答案
- NB-T20307-2014核電廠冷卻塔環境影響評價技術規范
- 高中數學選修二(人教A版2019)課后習題答案解析
- 天然氣管網大數據分析與預測
- 2024年中華人民共和國企業所得稅年度納稅申報表(帶公式)20240301更新
- 蔬菜購銷合同電子版模板
- 創新創業基礎-理論、案例與訓練(大學生創新創業教育課程)全套教學課件
- 公廁保潔服務服務承諾及質量保障措施
評論
0/150
提交評論