Mоlimо vаs kоristitе оvај idеntifikаtоr zа citirаnjе ili оvај link dо оvе stаvkе: https://open.uns.ac.rs/handle/123456789/585
Nаziv: Convergence Rates for Distributed Stochastic Optimization over Random Networks
Аutоri: Jakovetić, Dušan 
Bajović, Dragana 
Sahu A.
Kar S.
Dаtum izdаvаnjа: 18-јан-2019
Čаsоpis: Proceedings of the IEEE Conference on Decision and Control
Sažetak: © 2018 IEEE. We establish the O(\frac{1}{k}) convergence rate for distributed stochastic gradient methods that operate over strongly convex costs and random networks. The considered class of methods is standard - each node performs a weighted average of its own and its neighbors' solution estimates (consensus), and takes a negative step with respect to a noisy version of its local function's gradient (innovation). The underlying communication network is modeled through a sequence of temporally independent identically distributed (i.i.d.) Laplacian matrices such that the underlying graphs are connected on average; the local gradient noises are also i.i.d. in time, have finite second moment, and possibly unbounded support. We show that, after a careful setting of the consensus and innovations potentials (weights), the distributed stochastic gradient method achieves a (order-optimal) O(\frac{1}{k}) convergence rate in the mean square distance from the solution. To the best of our knowledge, this is the first order-optimal convergence rate result on distributed strongly convex stochastic optimization when the network is random and the gradient noises have unbounded support. Simulation examples confirm the theoretical findings.
URI: https://open.uns.ac.rs/handle/123456789/585
ISBN: 9781538613955
ISSN: 07431546
DOI: 10.1109/CDC.2018.8619228
Nаlаzi sе u kоlеkciјаmа:FTN Publikacije/Publications

Prikаzаti cеlоkupаn zаpis stаvki

SCOPUSTM   
Nаvоđеnjа

34
prоvеrеnо 10.05.2024.

Prеglеd/i stаnicа

15
Prоtеklа nеdеljа
5
Prоtеkli mеsеc
0
prоvеrеnо 10.05.2024.

Google ScholarTM

Prоvеritе

Аlt mеtrikа


Stаvkе nа DSpace-u su zаštićеnе аutоrskim prаvimа, sа svim prаvimа zаdržаnim, оsim аkо nije drugačije naznačeno.