1
#![cfg_attr(docsrs, feature(doc_cfg))]
2
#![doc = include_str!("../README.md")]
3
// @@ begin lint list maintained by maint/add_warning @@
4
#![allow(renamed_and_removed_lints)] // @@REMOVE_WHEN(ci_arti_stable)
5
#![allow(unknown_lints)] // @@REMOVE_WHEN(ci_arti_nightly)
6
#![warn(missing_docs)]
7
#![warn(noop_method_call)]
8
#![warn(unreachable_pub)]
9
#![warn(clippy::all)]
10
#![deny(clippy::await_holding_lock)]
11
#![deny(clippy::cargo_common_metadata)]
12
#![deny(clippy::cast_lossless)]
13
#![deny(clippy::checked_conversions)]
14
#![warn(clippy::cognitive_complexity)]
15
#![deny(clippy::debug_assert_with_mut_call)]
16
#![deny(clippy::exhaustive_enums)]
17
#![deny(clippy::exhaustive_structs)]
18
#![deny(clippy::expl_impl_clone_on_copy)]
19
#![deny(clippy::fallible_impl_from)]
20
#![deny(clippy::implicit_clone)]
21
#![deny(clippy::large_stack_arrays)]
22
#![warn(clippy::manual_ok_or)]
23
#![deny(clippy::missing_docs_in_private_items)]
24
#![warn(clippy::needless_borrow)]
25
#![warn(clippy::needless_pass_by_value)]
26
#![warn(clippy::option_option)]
27
#![deny(clippy::print_stderr)]
28
#![deny(clippy::print_stdout)]
29
#![warn(clippy::rc_buffer)]
30
#![deny(clippy::ref_option_ref)]
31
#![warn(clippy::semicolon_if_nothing_returned)]
32
#![warn(clippy::trait_duplication_in_bounds)]
33
#![deny(clippy::unchecked_time_subtraction)]
34
#![deny(clippy::unnecessary_wraps)]
35
#![warn(clippy::unseparated_literal_suffix)]
36
#![deny(clippy::unwrap_used)]
37
#![deny(clippy::mod_module_files)]
38
#![allow(clippy::let_unit_value)] // This can reasonably be done for explicitness
39
#![allow(clippy::uninlined_format_args)]
40
#![allow(clippy::significant_drop_in_scrutinee)] // arti/-/merge_requests/588/#note_2812945
41
#![allow(clippy::result_large_err)] // temporary workaround for arti#587
42
#![allow(clippy::needless_raw_string_hashes)] // complained-about code is fine, often best
43
#![allow(clippy::needless_lifetimes)] // See arti#1765
44
#![allow(mismatched_lifetime_syntaxes)] // temporary workaround for arti#2060
45
#![allow(clippy::collapsible_if)] // See arti#2342
46
#![deny(clippy::unused_async)]
47
//! <!-- @@ end lint list maintained by maint/add_warning @@ -->
48

            
49
pub mod details;
50
mod err;
51
#[cfg(feature = "hs-common")]
52
mod hsdir_params;
53
#[cfg(feature = "hs-common")]
54
mod hsdir_ring;
55
pub mod params;
56
mod weight;
57

            
58
#[cfg(any(test, feature = "testing"))]
59
pub mod testnet;
60
#[cfg(feature = "testing")]
61
pub mod testprovider;
62

            
63
use async_trait::async_trait;
64
#[cfg(feature = "hs-service")]
65
use itertools::chain;
66
use tor_error::warn_report;
67
#[cfg(feature = "hs-common")]
68
use tor_linkspec::OwnedCircTarget;
69
use tor_linkspec::{
70
    ChanTarget, DirectChanMethodsHelper, HasAddrs, HasRelayIds, RelayIdRef, RelayIdType,
71
};
72
use tor_llcrypto as ll;
73
use tor_llcrypto::pk::{ed25519::Ed25519Identity, rsa::RsaIdentity};
74
use tor_netdoc::doc::microdesc::{MdDigest, Microdesc};
75
use tor_netdoc::doc::netstatus::{self, MdConsensus, MdRouterStatus};
76
#[cfg(feature = "hs-common")]
77
use {hsdir_ring::HsDirRing, std::iter};
78

            
79
use derive_more::{From, Into};
80
use futures::{StreamExt, stream::BoxStream};
81
use num_enum::{IntoPrimitive, TryFromPrimitive};
82
use rand::seq::{IndexedRandom as _, SliceRandom as _, WeightError};
83
use serde::Deserialize;
84
use std::collections::HashMap;
85
use std::net::IpAddr;
86
use std::ops::Deref;
87
use std::sync::Arc;
88
use std::time::SystemTime;
89
use strum::{EnumCount, EnumIter};
90
use tracing::warn;
91
use typed_index_collections::{TiSlice, TiVec};
92

            
93
#[cfg(feature = "hs-common")]
94
use {
95
    itertools::Itertools,
96
    std::collections::HashSet,
97
    std::result::Result as StdResult,
98
    tor_error::{Bug, internal},
99
    tor_hscrypto::{pk::HsBlindId, time::TimePeriod},
100
    tor_linkspec::{OwnedChanTargetBuilder, verbatim::VerbatimLinkSpecCircTarget},
101
    tor_llcrypto::pk::curve25519,
102
};
103

            
104
pub use err::Error;
105
pub use weight::WeightRole;
106
/// A Result using the Error type from the tor-netdir crate
107
pub type Result<T> = std::result::Result<T, Error>;
108

            
109
#[cfg(feature = "hs-common")]
110
pub use err::{OnionDirLookupError, VerbatimCircTargetDecodeError};
111

            
112
use params::NetParameters;
113
#[cfg(feature = "geoip")]
114
use tor_geoip::{CountryCode, GeoipDb, HasCountryCode};
115

            
116
#[cfg(feature = "hs-common")]
117
pub use hsdir_params::HsDirParams;
118

            
119
/// Index into the consensus relays
120
///
121
/// This is an index into the list of relays returned by
122
/// [`.c_relays()`](ConsensusRelays::c_relays)
123
/// (on the corresponding consensus or netdir).
124
///
125
/// This is just a `usize` inside, but using a newtype prevents getting a relay index
126
/// confused with other kinds of slice indices or counts.
127
///
128
/// If you are in a part of the code which needs to work with multiple consensuses,
129
/// the typechecking cannot tell if you try to index into the wrong consensus.
130
#[derive(Debug, From, Into, Copy, Clone, Ord, PartialOrd, Eq, PartialEq, Hash)]
131
pub(crate) struct RouterStatusIdx(usize);
132

            
133
/// Extension trait to provide index-type-safe `.c_relays()` method
134
//
135
// TODO: Really it would be better to have MdConsensns::relays() return TiSlice,
136
// but that would be an API break there.
137
pub(crate) trait ConsensusRelays {
138
    /// Obtain the list of relays in the consensus
139
    //
140
    fn c_relays(&self) -> &TiSlice<RouterStatusIdx, MdRouterStatus>;
141
}
142
impl ConsensusRelays for MdConsensus {
143
53508357
    fn c_relays(&self) -> &TiSlice<RouterStatusIdx, MdRouterStatus> {
144
53508357
        TiSlice::from_ref(MdConsensus::relays(self))
145
53508357
    }
146
}
147
impl ConsensusRelays for NetDir {
148
53410060
    fn c_relays(&self) -> &TiSlice<RouterStatusIdx, MdRouterStatus> {
149
53410060
        self.consensus.c_relays()
150
53410060
    }
151
}
152

            
153
/// Configuration for determining when two relays have addresses "too close" in
154
/// the network.
155
///
156
/// Used by `Relay::low_level_details().in_same_subnet()`.
157
#[derive(Deserialize, Debug, Clone, Copy, Eq, PartialEq)]
158
#[serde(deny_unknown_fields)]
159
pub struct SubnetConfig {
160
    /// Consider IPv4 nodes in the same /x to be the same family.
161
    ///
162
    /// If this value is 0, all nodes with IPv4 addresses will be in the
163
    /// same family.  If this value is above 32, then no nodes will be
164
    /// placed im the same family based on their IPv4 addresses.
165
    subnets_family_v4: u8,
166
    /// Consider IPv6 nodes in the same /x to be the same family.
167
    ///
168
    /// If this value is 0, all nodes with IPv6 addresses will be in the
169
    /// same family.  If this value is above 128, then no nodes will be
170
    /// placed im the same family based on their IPv6 addresses.
171
    subnets_family_v6: u8,
172
}
173

            
174
impl Default for SubnetConfig {
175
570161
    fn default() -> Self {
176
570161
        Self::new(16, 32)
177
570161
    }
178
}
179

            
180
impl SubnetConfig {
181
    /// Construct a new SubnetConfig from a pair of bit prefix lengths.
182
    ///
183
    /// The values are clamped to the appropriate ranges if they are
184
    /// out-of-bounds.
185
1731954
    pub fn new(subnets_family_v4: u8, subnets_family_v6: u8) -> Self {
186
1731954
        Self {
187
1731954
            subnets_family_v4,
188
1731954
            subnets_family_v6,
189
1731954
        }
190
1731954
    }
191

            
192
    /// Construct a new SubnetConfig such that addresses are not in the same
193
    /// family with anything--not even with themselves.
194
3488718
    pub fn no_addresses_match() -> SubnetConfig {
195
3488718
        SubnetConfig {
196
3488718
            subnets_family_v4: 33,
197
3488718
            subnets_family_v6: 129,
198
3488718
        }
199
3488718
    }
200

            
201
    /// Return true if the two addresses in the same subnet, according to this
202
    /// configuration.
203
48904320
    pub fn addrs_in_same_subnet(&self, a: &IpAddr, b: &IpAddr) -> bool {
204
48904320
        match (a, b) {
205
48904288
            (IpAddr::V4(a), IpAddr::V4(b)) => {
206
48904288
                let bits = self.subnets_family_v4;
207
48904288
                if bits > 32 {
208
6964
                    return false;
209
48897324
                }
210
48897324
                let a = u32::from_be_bytes(a.octets());
211
48897324
                let b = u32::from_be_bytes(b.octets());
212
48897324
                (a >> (32 - bits)) == (b >> (32 - bits))
213
            }
214
20
            (IpAddr::V6(a), IpAddr::V6(b)) => {
215
20
                let bits = self.subnets_family_v6;
216
20
                if bits > 128 {
217
4
                    return false;
218
16
                }
219
16
                let a = u128::from_be_bytes(a.octets());
220
16
                let b = u128::from_be_bytes(b.octets());
221
16
                (a >> (128 - bits)) == (b >> (128 - bits))
222
            }
223
12
            _ => false,
224
        }
225
48904320
    }
226

            
227
    /// Return true if any of the addresses in `a` shares a subnet with any of
228
    /// the addresses in `b`, according to this configuration.
229
48904270
    pub fn any_addrs_in_same_subnet<T, U>(&self, a: &T, b: &U) -> bool
230
48904270
    where
231
48904270
        T: tor_linkspec::HasAddrs,
232
48904270
        U: tor_linkspec::HasAddrs,
233
    {
234
48904276
        a.addrs().any(|aa| {
235
48904276
            b.addrs()
236
48904284
                .any(|bb| self.addrs_in_same_subnet(&aa.ip(), &bb.ip()))
237
48904276
        })
238
48904270
    }
239

            
240
    /// Return a new subnet configuration that is the union of `self` and
241
    /// `other`.
242
    ///
243
    /// That is, return a subnet configuration that puts all addresses in the
244
    /// same subnet if and only if at least one of `self` and `other` would put
245
    /// them in the same subnet.
246
1741262
    pub fn union(&self, other: &Self) -> Self {
247
        use std::cmp::min;
248
1741262
        Self {
249
1741262
            subnets_family_v4: min(self.subnets_family_v4, other.subnets_family_v4),
250
1741262
            subnets_family_v6: min(self.subnets_family_v6, other.subnets_family_v6),
251
1741262
        }
252
1741262
    }
253
}
254

            
255
/// Configuration for which listed family information to use when deciding
256
/// whether relays belong to the same family.
257
///
258
/// Derived from network parameters.
259
#[derive(Clone, Copy, Debug)]
260
pub struct FamilyRules {
261
    /// If true, we use family information from lists of family members.
262
    use_family_lists: bool,
263
    /// If true, we use family information from lists of family IDs and from family certs.
264
    use_family_ids: bool,
265
}
266

            
267
impl<'a> From<&'a NetParameters> for FamilyRules {
268
580826
    fn from(params: &'a NetParameters) -> Self {
269
580826
        FamilyRules {
270
580826
            use_family_lists: bool::from(params.use_family_lists),
271
580826
            use_family_ids: bool::from(params.use_family_ids),
272
580826
        }
273
580826
    }
274
}
275

            
276
impl FamilyRules {
277
    /// Return a `FamilyRules` that will use all recognized kinds of family information.
278
895
    pub fn all_family_info() -> Self {
279
895
        Self {
280
895
            use_family_lists: true,
281
895
            use_family_ids: true,
282
895
        }
283
895
    }
284

            
285
    /// Return a `FamilyRules` that will ignore all family information declared by relays.
286
3488810
    pub fn ignore_declared_families() -> Self {
287
3488810
        Self {
288
3488810
            use_family_lists: false,
289
3488810
            use_family_ids: false,
290
3488810
        }
291
3488810
    }
292

            
293
    /// Configure this `FamilyRules` to use (or not use) family information from
294
    /// lists of family members.
295
47
    pub fn use_family_lists(&mut self, val: bool) -> &mut Self {
296
47
        self.use_family_lists = val;
297
47
        self
298
47
    }
299

            
300
    /// Configure this `FamilyRules` to use (or not use) family information from
301
    /// family IDs and family certs.
302
47
    pub fn use_family_ids(&mut self, val: bool) -> &mut Self {
303
47
        self.use_family_ids = val;
304
47
        self
305
47
    }
306

            
307
    /// Return a `FamilyRules` that will look at every source of information
308
    /// requested by `self` or by `other`.
309
1741256
    pub fn union(&self, other: &Self) -> Self {
310
        Self {
311
1741256
            use_family_lists: self.use_family_lists || other.use_family_lists,
312
1741256
            use_family_ids: self.use_family_ids || other.use_family_ids,
313
        }
314
1741256
    }
315
}
316

            
317
/// An opaque type representing the weight with which a relay or set of
318
/// relays will be selected for a given role.
319
///
320
/// Most users should ignore this type, and just use pick_relay instead.
321
#[derive(
322
    Copy,
323
    Clone,
324
    Debug,
325
    derive_more::Add,
326
    derive_more::Sum,
327
    derive_more::AddAssign,
328
    Eq,
329
    PartialEq,
330
    Ord,
331
    PartialOrd,
332
)]
333
pub struct RelayWeight(u64);
334

            
335
impl RelayWeight {
336
    /// Try to divide this weight by `rhs`.
337
    ///
338
    /// Return a ratio on success, or None on division-by-zero.
339
3529
    pub fn checked_div(&self, rhs: RelayWeight) -> Option<f64> {
340
3529
        if rhs.0 == 0 {
341
2
            None
342
        } else {
343
3527
            Some((self.0 as f64) / (rhs.0 as f64))
344
        }
345
3529
    }
346

            
347
    /// Compute a ratio `frac` of this weight.
348
    ///
349
    /// Return None if frac is less than zero, since negative weights
350
    /// are impossible.
351
27217
    pub fn ratio(&self, frac: f64) -> Option<RelayWeight> {
352
27217
        let product = (self.0 as f64) * frac;
353
27217
        if product >= 0.0 && product.is_finite() {
354
27215
            Some(RelayWeight(product as u64))
355
        } else {
356
2
            None
357
        }
358
27217
    }
359
}
360

            
361
impl From<u64> for RelayWeight {
362
7050
    fn from(val: u64) -> Self {
363
7050
        RelayWeight(val)
364
7050
    }
365
}
366

            
367
/// An operation for which we might be requesting a hidden service directory.
368
#[derive(Copy, Clone, Debug, PartialEq)]
369
// TODO: make this pub(crate) once NetDir::hs_dirs is removed
370
#[non_exhaustive]
371
pub enum HsDirOp {
372
    /// Uploading an onion service descriptor.
373
    #[cfg(feature = "hs-service")]
374
    Upload,
375
    /// Downloading an onion service descriptor.
376
    Download,
377
}
378

            
379
/// A view of the Tor directory, suitable for use in building circuits.
380
///
381
/// Abstractly, a [`NetDir`] is a set of usable public [`Relay`]s, each of which
382
/// has its own properties, identity, and correct weighted probability for use
383
/// under different circumstances.
384
///
385
/// A [`NetDir`] is constructed by making a [`PartialNetDir`] from a consensus
386
/// document, and then adding enough microdescriptors to that `PartialNetDir` so
387
/// that it can be used to build paths. (Thus, if you have a NetDir, it is
388
/// definitely adequate to build paths.)
389
///
390
/// # "Usable" relays
391
///
392
/// Many methods on NetDir are defined in terms of <a name="usable">"Usable"</a> relays.  Unless
393
/// otherwise stated, a relay is "usable" if it is listed in the consensus,
394
/// if we have full directory information for that relay (including a
395
/// microdescriptor), and if that relay does not have any flags indicating that
396
/// we should never use it. (Currently, `NoEdConsensus` is the only such flag.)
397
///
398
/// # Limitations
399
///
400
/// The current NetDir implementation assumes fairly strongly that every relay
401
/// has an Ed25519 identity and an RSA identity, that the consensus is indexed
402
/// by RSA identities, and that the Ed25519 identities are stored in
403
/// microdescriptors.
404
///
405
/// If these assumptions someday change, then we'll have to revise the
406
/// implementation.
407
#[derive(Debug, Clone)]
408
pub struct NetDir {
409
    /// A microdescriptor consensus that lists the members of the network,
410
    /// and maps each one to a 'microdescriptor' that has more information
411
    /// about it
412
    consensus: Arc<MdConsensus>,
413
    /// A map from keys to integer values, distributed in the consensus,
414
    /// and clamped to certain defaults.
415
    params: NetParameters,
416
    /// Map from routerstatus index, to that routerstatus's microdescriptor (if we have one.)
417
    mds: TiVec<RouterStatusIdx, Option<Arc<Microdesc>>>,
418
    /// Map from SHA256 of _missing_ microdescriptors to the index of their
419
    /// corresponding routerstatus.
420
    rsidx_by_missing: HashMap<MdDigest, RouterStatusIdx>,
421
    /// Map from ed25519 identity to index of the routerstatus.
422
    ///
423
    /// Note that we don't know the ed25519 identity of a relay until
424
    /// we get the microdescriptor for it, so this won't be filled in
425
    /// until we get the microdescriptors.
426
    ///
427
    /// # Implementation note
428
    ///
429
    /// For this field, and for `rsidx_by_rsa`,
430
    /// it might be cool to have references instead.
431
    /// But that would make this into a self-referential structure,
432
    /// which isn't possible in safe rust.
433
    rsidx_by_ed: HashMap<Ed25519Identity, RouterStatusIdx>,
434
    /// Map from RSA identity to index of the routerstatus.
435
    ///
436
    /// This is constructed at the same time as the NetDir object, so it
437
    /// can be immutable.
438
    rsidx_by_rsa: Arc<HashMap<RsaIdentity, RouterStatusIdx>>,
439

            
440
    /// Hash ring(s) describing the onion service directory.
441
    ///
442
    /// This is empty in a PartialNetDir, and is filled in before the NetDir is
443
    /// built.
444
    //
445
    // TODO hs: It is ugly to have this exist in a partially constructed state
446
    // in a PartialNetDir.
447
    // Ideally, a PartialNetDir would contain only an HsDirs<HsDirParams>,
448
    // or perhaps nothing at all, here.
449
    #[cfg(feature = "hs-common")]
450
    hsdir_rings: Arc<HsDirs<HsDirRing>>,
451

            
452
    /// Weight values to apply to a given relay when deciding how frequently
453
    /// to choose it for a given role.
454
    weights: weight::WeightSet,
455

            
456
    #[cfg(feature = "geoip")]
457
    /// Country codes for each router in our consensus.
458
    ///
459
    /// This is indexed by the `RouterStatusIdx` (i.e. a router idx of zero has
460
    /// the country code at position zero in this array).
461
    country_codes: Vec<Option<CountryCode>>,
462
}
463

            
464
/// Collection of hidden service directories (or parameters for them)
465
///
466
/// In [`NetDir`] this is used to store the actual hash rings.
467
/// (But, in a NetDir in a [`PartialNetDir`], it contains [`HsDirRing`]s
468
/// where only the `params` are populated, and the `ring` is empty.)
469
///
470
/// This same generic type is used as the return type from
471
/// [`HsDirParams::compute`](HsDirParams::compute),
472
/// where it contains the *parameters* for the primary and secondary rings.
473
#[derive(Debug, Clone)]
474
#[cfg(feature = "hs-common")]
475
pub(crate) struct HsDirs<D> {
476
    /// The current ring
477
    ///
478
    /// It corresponds to the time period containing the `valid-after` time in
479
    /// the consensus. Its SRV is whatever SRV was most current at the time when
480
    /// that time period began.
481
    ///
482
    /// This is the hash ring that we should use whenever we are fetching an
483
    /// onion service descriptor.
484
    current: D,
485

            
486
    /// Secondary rings (based on the parameters for the previous and next time periods)
487
    ///
488
    /// Onion services upload to positions on these ring as well, based on how
489
    /// far into the current time period this directory is, so that
490
    /// not-synchronized clients can still find their descriptor.
491
    ///
492
    /// Note that with the current (2023) network parameters, with
493
    /// `hsdir_interval = SRV lifetime = 24 hours` at most one of these
494
    /// secondary rings will be active at a time.  We have two here in order
495
    /// to conform with a more flexible regime in proposal 342.
496
    //
497
    // TODO: hs clients never need this; so I've made it not-present for them.
498
    // But does that risk too much with respect to side channels?
499
    //
500
    // TODO: Perhaps we should refactor this so that it is clear that these
501
    // are immutable?  On the other hand, the documentation for this type
502
    // declares that it is immutable, so we are likely okay.
503
    //
504
    // TODO: this `Vec` is only ever 0,1,2 elements.
505
    // Maybe it should be an ArrayVec or something.
506
    #[cfg(feature = "hs-service")]
507
    secondary: Vec<D>,
508
}
509

            
510
#[cfg(feature = "hs-common")]
511
impl<D> HsDirs<D> {
512
    /// Convert an `HsDirs<D>` to `HsDirs<D2>` by mapping each contained `D`
513
19524
    pub(crate) fn map<D2>(self, mut f: impl FnMut(D) -> D2) -> HsDirs<D2> {
514
19524
        HsDirs {
515
19524
            current: f(self.current),
516
19524
            #[cfg(feature = "hs-service")]
517
19524
            secondary: self.secondary.into_iter().map(f).collect(),
518
19524
        }
519
19524
    }
520

            
521
    /// Iterate over some of the contained hsdirs, according to `secondary`
522
    ///
523
    /// The current ring is always included.
524
    /// Secondary rings are included iff `secondary` and the `hs-service` feature is enabled.
525
756
    fn iter_filter_secondary(&self, secondary: bool) -> impl Iterator<Item = &D> {
526
756
        let i = iter::once(&self.current);
527

            
528
        // With "hs-service" disabled, there are no secondary rings,
529
        // so we don't care.
530
756
        let _ = secondary;
531

            
532
        #[cfg(feature = "hs-service")]
533
756
        let i = chain!(i, self.secondary.iter().filter(move |_| secondary));
534

            
535
756
        i
536
756
    }
537

            
538
    /// Iterate over all the contained hsdirs
539
752
    pub(crate) fn iter(&self) -> impl Iterator<Item = &D> {
540
752
        self.iter_filter_secondary(true)
541
752
    }
542

            
543
    /// Iterate over the hsdirs relevant for `op`
544
4
    pub(crate) fn iter_for_op(&self, op: HsDirOp) -> impl Iterator<Item = &D> {
545
4
        self.iter_filter_secondary(match op {
546
            #[cfg(feature = "hs-service")]
547
2
            HsDirOp::Upload => true,
548
2
            HsDirOp::Download => false,
549
        })
550
4
    }
551
}
552

            
553
/// An event that a [`NetDirProvider`] can broadcast to indicate that a change in
554
/// the status of its directory.
555
#[derive(
556
    Debug, Clone, Copy, PartialEq, Eq, EnumIter, EnumCount, IntoPrimitive, TryFromPrimitive,
557
)]
558
#[non_exhaustive]
559
#[repr(u16)]
560
pub enum DirEvent {
561
    /// A new consensus has been received, and has enough information to be
562
    /// used.
563
    ///
564
    /// This event is also broadcast when a new set of consensus parameters is
565
    /// available, even if that set of parameters comes from a configuration
566
    /// change rather than from the latest consensus.
567
    NewConsensus,
568

            
569
    /// New descriptors have been received for the current consensus.
570
    ///
571
    /// (This event is _not_ broadcast when receiving new descriptors for a
572
    /// consensus which is not yet ready to replace the current consensus.)
573
    NewDescriptors,
574

            
575
    /// We have received updated recommendations and requirements
576
    /// for which subprotocols we should have to use the network.
577
    NewProtocolRecommendation,
578
}
579

            
580
/// The network directory provider is shutting down without giving us the
581
/// netdir we asked for.
582
#[derive(Clone, Copy, Debug, thiserror::Error)]
583
#[error("Network directory provider is shutting down")]
584
#[non_exhaustive]
585
pub struct NetdirProviderShutdown;
586

            
587
impl tor_error::HasKind for NetdirProviderShutdown {
588
    fn kind(&self) -> tor_error::ErrorKind {
589
        tor_error::ErrorKind::ArtiShuttingDown
590
    }
591
}
592

            
593
/// How "timely" must a network directory be?
594
///
595
/// This enum is used as an argument when requesting a [`NetDir`] object from
596
/// [`NetDirProvider`] and other APIs, to specify how recent the information
597
/// must be in order to be useful.
598
#[derive(Copy, Clone, Eq, PartialEq, Debug)]
599
#[allow(clippy::exhaustive_enums)]
600
pub enum Timeliness {
601
    /// The network directory must be strictly timely.
602
    ///
603
    /// That is, it must be based on a consensus that valid right now, with no
604
    /// tolerance for skew or consensus problems.
605
    ///
606
    /// Avoid using this option if you could use [`Timeliness::Timely`] instead.
607
    Strict,
608
    /// The network directory must be roughly timely.
609
    ///
610
    /// This is, it must be be based on a consensus that is not _too_ far in the
611
    /// future, and not _too_ far in the past.
612
    ///
613
    /// (The tolerances for "too far" will depend on configuration.)
614
    ///
615
    /// This is almost always the option that you want to use.
616
    Timely,
617
    /// Any network directory is permissible, regardless of how untimely.
618
    ///
619
    /// Avoid using this option if you could use [`Timeliness::Timely`] instead.
620
    Unchecked,
621
}
622

            
623
/// An object that can provide [`NetDir`]s, as well as inform consumers when
624
/// they might have changed.
625
///
626
/// It is the responsibility of the implementor of `NetDirProvider`
627
/// to try to obtain an up-to-date `NetDir`,
628
/// and continuously to maintain and update it.
629
///
630
/// In usual configurations, Arti uses `tor_dirmgr::DirMgr`
631
/// as its `NetDirProvider`.
632
#[async_trait]
633
pub trait NetDirProvider: UpcastArcNetDirProvider + Send + Sync {
634
    /// Return a network directory that's live according to the provided
635
    /// `timeliness`.
636
    fn netdir(&self, timeliness: Timeliness) -> Result<Arc<NetDir>>;
637

            
638
    /// Return a reasonable netdir for general usage.
639
    ///
640
    /// This is an alias for
641
    /// [`NetDirProvider::netdir`]`(`[`Timeliness::Timely`]`)`.
642
792
    fn timely_netdir(&self) -> Result<Arc<NetDir>> {
643
792
        self.netdir(Timeliness::Timely)
644
792
    }
645

            
646
    /// Return a new asynchronous stream that will receive notification
647
    /// whenever the consensus has changed.
648
    ///
649
    /// Multiple events may be batched up into a single item: each time
650
    /// this stream yields an event, all you can assume is that the event has
651
    /// occurred at least once.
652
    fn events(&self) -> BoxStream<'static, DirEvent>;
653

            
654
    /// Return the latest network parameters.
655
    ///
656
    /// If we have no directory, return a reasonable set of defaults.
657
    fn params(&self) -> Arc<dyn AsRef<NetParameters>>;
658

            
659
    /// Get a NetDir from `provider`, waiting until one exists.
660
    async fn wait_for_netdir(
661
        &self,
662
        timeliness: Timeliness,
663
64
    ) -> std::result::Result<Arc<NetDir>, NetdirProviderShutdown> {
664
        if let Ok(nd) = self.netdir(timeliness) {
665
            return Ok(nd);
666
        }
667

            
668
        let mut stream = self.events();
669
        loop {
670
            // We need to retry `self.netdir()` before waiting for any stream events, to
671
            // avoid deadlock.
672
            //
673
            // We ignore all errors here: they can all potentially be fixed by
674
            // getting a fresh consensus, and they will all get warned about
675
            // by the NetDirProvider itself.
676
            if let Ok(nd) = self.netdir(timeliness) {
677
                return Ok(nd);
678
            }
679
            match stream.next().await {
680
                Some(_) => {}
681
                None => {
682
                    return Err(NetdirProviderShutdown);
683
                }
684
            }
685
        }
686
64
    }
687

            
688
    /// Wait until `provider` lists `target`.
689
    ///
690
    /// NOTE: This might potentially wait indefinitely, if `target` is never actually
691
    /// becomes listed in the directory.  It will exit if the `NetDirProvider` shuts down.
692
    async fn wait_for_netdir_to_list(
693
        &self,
694
        target: &tor_linkspec::RelayIds,
695
        timeliness: Timeliness,
696
    ) -> std::result::Result<(), NetdirProviderShutdown> {
697
        let mut events = self.events();
698
        loop {
699
            // See if the desired relay is in the netdir.
700
            //
701
            // We do this before waiting for any events, to avoid race conditions.
702
            {
703
                let netdir = self.wait_for_netdir(timeliness).await?;
704
                if netdir.ids_listed(target) == Some(true) {
705
                    return Ok(());
706
                }
707
                // If we reach this point, then ids_listed returned `Some(false)`,
708
                // meaning "This relay is definitely not in the current directory";
709
                // or it returned `None`, meaning "waiting for more information
710
                // about this network directory.
711
                // In both cases, it's reasonable to just wait for another netdir
712
                // event and try again.
713
            }
714
            // We didn't find the relay; wait for the provider to have a new netdir
715
            // or more netdir information.
716
            if events.next().await.is_none() {
717
                // The event stream is closed; the provider has shut down.
718
                return Err(NetdirProviderShutdown);
719
            }
720
        }
721
    }
722

            
723
    /// Return the latest set of recommended and required protocols, if there is one.
724
    ///
725
    /// This may be more recent (or more available) than this provider's associated NetDir.
726
    fn protocol_statuses(&self) -> Option<(SystemTime, Arc<netstatus::ProtoStatuses>)>;
727
}
728

            
729
impl<T> NetDirProvider for Arc<T>
730
where
731
    T: NetDirProvider,
732
{
733
128
    fn netdir(&self, timeliness: Timeliness) -> Result<Arc<NetDir>> {
734
128
        self.deref().netdir(timeliness)
735
128
    }
736

            
737
16
    fn timely_netdir(&self) -> Result<Arc<NetDir>> {
738
16
        self.deref().timely_netdir()
739
16
    }
740

            
741
190
    fn events(&self) -> BoxStream<'static, DirEvent> {
742
190
        self.deref().events()
743
190
    }
744

            
745
26
    fn params(&self) -> Arc<dyn AsRef<NetParameters>> {
746
26
        self.deref().params()
747
26
    }
748

            
749
22
    fn protocol_statuses(&self) -> Option<(SystemTime, Arc<netstatus::ProtoStatuses>)> {
750
22
        self.deref().protocol_statuses()
751
22
    }
752
}
753

            
754
/// Helper trait: allows any `Arc<X>` to be upcast to a `Arc<dyn
755
/// NetDirProvider>` if X is an implementation or supertrait of NetDirProvider.
756
///
757
/// This trait exists to work around a limitation in rust: when trait upcasting
758
/// coercion is stable, this will be unnecessary.
759
///
760
/// The Rust tracking issue is <https://github.com/rust-lang/rust/issues/65991>.
761
pub trait UpcastArcNetDirProvider {
762
    /// Return a view of this object as an `Arc<dyn NetDirProvider>`
763
    fn upcast_arc<'a>(self: Arc<Self>) -> Arc<dyn NetDirProvider + 'a>
764
    where
765
        Self: 'a;
766
}
767

            
768
impl<T> UpcastArcNetDirProvider for T
769
where
770
    T: NetDirProvider + Sized,
771
{
772
118
    fn upcast_arc<'a>(self: Arc<Self>) -> Arc<dyn NetDirProvider + 'a>
773
118
    where
774
118
        Self: 'a,
775
    {
776
118
        self
777
118
    }
778
}
779

            
780
impl AsRef<NetParameters> for NetDir {
781
1598
    fn as_ref(&self) -> &NetParameters {
782
1598
        self.params()
783
1598
    }
784
}
785

            
786
/// A partially build NetDir -- it can't be unwrapped until it has
787
/// enough information to build safe paths.
788
#[derive(Debug, Clone)]
789
pub struct PartialNetDir {
790
    /// The netdir that's under construction.
791
    netdir: NetDir,
792

            
793
    /// The previous netdir, if we had one
794
    ///
795
    /// Used as a cache, so we can reuse information
796
    #[cfg(feature = "hs-common")]
797
    prev_netdir: Option<Arc<NetDir>>,
798
}
799

            
800
/// A view of a relay on the Tor network, suitable for building circuits.
801
// TODO: This should probably be a more specific struct, with a trait
802
// that implements it.
803
#[derive(Clone)]
804
pub struct Relay<'a> {
805
    /// A router descriptor for this relay.
806
    rs: &'a netstatus::MdRouterStatus,
807
    /// A microdescriptor for this relay.
808
    md: &'a Microdesc,
809
    /// The country code this relay is in, if we know one.
810
    #[cfg(feature = "geoip")]
811
    cc: Option<CountryCode>,
812
}
813

            
814
/// A relay that we haven't checked for validity or usability in
815
/// routing.
816
#[derive(Debug)]
817
pub struct UncheckedRelay<'a> {
818
    /// A router descriptor for this relay.
819
    rs: &'a netstatus::MdRouterStatus,
820
    /// A microdescriptor for this relay, if there is one.
821
    md: Option<&'a Microdesc>,
822
    /// The country code this relay is in, if we know one.
823
    #[cfg(feature = "geoip")]
824
    cc: Option<CountryCode>,
825
}
826

            
827
/// A partial or full network directory that we can download
828
/// microdescriptors for.
829
pub trait MdReceiver {
830
    /// Return an iterator over the digests for all of the microdescriptors
831
    /// that this netdir is missing.
832
    fn missing_microdescs(&self) -> Box<dyn Iterator<Item = &MdDigest> + '_>;
833
    /// Add a microdescriptor to this netdir, if it was wanted.
834
    ///
835
    /// Return true if it was indeed wanted.
836
    fn add_microdesc(&mut self, md: Microdesc) -> bool;
837
    /// Return the number of missing microdescriptors.
838
    fn n_missing(&self) -> usize;
839
}
840

            
841
impl PartialNetDir {
842
    /// Create a new PartialNetDir with a given consensus, and no
843
    /// microdescriptors loaded.
844
    ///
845
    /// If `replacement_params` is provided, override network parameters from
846
    /// the consensus with those from `replacement_params`.
847
9671
    pub fn new(
848
9671
        consensus: MdConsensus,
849
9671
        replacement_params: Option<&netstatus::NetParams<i32>>,
850
9671
    ) -> Self {
851
9671
        Self::new_inner(
852
9671
            consensus,
853
9671
            replacement_params,
854
            #[cfg(feature = "geoip")]
855
9671
            None,
856
        )
857
9671
    }
858

            
859
    /// Create a new PartialNetDir with GeoIP support.
860
    ///
861
    /// This does the same thing as `new()`, except the provided GeoIP database is used to add
862
    /// country codes to relays.
863
    #[cfg(feature = "geoip")]
864
143
    pub fn new_with_geoip(
865
143
        consensus: MdConsensus,
866
143
        replacement_params: Option<&netstatus::NetParams<i32>>,
867
143
        geoip_db: &GeoipDb,
868
143
    ) -> Self {
869
143
        Self::new_inner(consensus, replacement_params, Some(geoip_db))
870
143
    }
871

            
872
    /// Implementation of the `new()` functions.
873
9814
    fn new_inner(
874
9814
        consensus: MdConsensus,
875
9814
        replacement_params: Option<&netstatus::NetParams<i32>>,
876
9814
        #[cfg(feature = "geoip")] geoip_db: Option<&GeoipDb>,
877
9814
    ) -> Self {
878
9814
        let mut params = NetParameters::default();
879

            
880
        // (We ignore unrecognized options here, since they come from
881
        // the consensus, and we don't expect to recognize everything
882
        // there.)
883
9814
        let _ = params.saturating_update(consensus.params().iter());
884

            
885
        // Now see if the user has any parameters to override.
886
        // (We have to do this now, or else changes won't be reflected in our
887
        // weights.)
888
9814
        if let Some(replacement) = replacement_params {
889
9802
            for u in params.saturating_update(replacement.iter()) {
890
2
                warn!("Unrecognized option: override_net_params.{}", u);
891
            }
892
12
        }
893

            
894
        // Compute the weights we'll want to use for these relays.
895
9814
        let weights = weight::WeightSet::from_consensus(&consensus, &params);
896

            
897
9814
        let n_relays = consensus.c_relays().len();
898

            
899
9814
        let rsidx_by_missing = consensus
900
9814
            .c_relays()
901
9814
            .iter_enumerated()
902
371402
            .map(|(rsidx, rs)| (*rs.md_digest(), rsidx))
903
9814
            .collect();
904

            
905
9814
        let rsidx_by_rsa = consensus
906
9814
            .c_relays()
907
9814
            .iter_enumerated()
908
371402
            .map(|(rsidx, rs)| (*rs.rsa_identity(), rsidx))
909
9814
            .collect();
910

            
911
        #[cfg(feature = "geoip")]
912
9814
        let country_codes = if let Some(db) = geoip_db {
913
143
            consensus
914
143
                .c_relays()
915
143
                .iter()
916
742
                .map(|rs| {
917
1030
                    db.lookup_country_code_multi(rs.addrs().map(|x| x.ip()))
918
738
                        .cloned()
919
738
                })
920
143
                .collect()
921
        } else {
922
9671
            Default::default()
923
        };
924

            
925
        #[cfg(feature = "hs-common")]
926
9814
        let hsdir_rings = Arc::new({
927
9814
            let params = HsDirParams::compute(&consensus, &params).expect("Invalid consensus!");
928
            // TODO: It's a bit ugly to use expect above, but this function does
929
            // not return a Result. On the other hand, the error conditions under which
930
            // HsDirParams::compute can return Err are _very_ narrow and hard to
931
            // hit; see documentation in that function.  As such, we probably
932
            // don't need to have this return a Result.
933

            
934
9814
            params.map(HsDirRing::empty_from_params)
935
        });
936

            
937
9814
        let netdir = NetDir {
938
9814
            consensus: Arc::new(consensus),
939
9814
            params,
940
9814
            mds: vec![None; n_relays].into(),
941
9814
            rsidx_by_missing,
942
9814
            rsidx_by_rsa: Arc::new(rsidx_by_rsa),
943
9814
            rsidx_by_ed: HashMap::with_capacity(n_relays),
944
9814
            #[cfg(feature = "hs-common")]
945
9814
            hsdir_rings,
946
9814
            weights,
947
9814
            #[cfg(feature = "geoip")]
948
9814
            country_codes,
949
9814
        };
950

            
951
9814
        PartialNetDir {
952
9814
            netdir,
953
9814
            #[cfg(feature = "hs-common")]
954
9814
            prev_netdir: None,
955
9814
        }
956
9814
    }
957

            
958
    /// Return the declared lifetime of this PartialNetDir.
959
2
    pub fn lifetime(&self) -> &netstatus::Lifetime {
960
2
        self.netdir.lifetime()
961
2
    }
962

            
963
    /// Record a previous netdir, which can be used for reusing cached information
964
    //
965
    // Fills in as many missing microdescriptors as possible in this
966
    // netdir, using the microdescriptors from the previous netdir.
967
    //
968
    // With HS enabled, stores the netdir for reuse of relay hash ring index values.
969
    #[allow(clippy::needless_pass_by_value)] // prev might, or might not, be stored
970
2
    pub fn fill_from_previous_netdir(&mut self, prev: Arc<NetDir>) {
971
76
        for md in prev.mds.iter().flatten() {
972
76
            self.netdir.add_arc_microdesc(md.clone());
973
76
        }
974

            
975
        #[cfg(feature = "hs-common")]
976
2
        {
977
2
            self.prev_netdir = Some(prev);
978
2
        }
979
2
    }
980

            
981
    /// Compute the hash ring(s) for this NetDir
982
    #[cfg(feature = "hs-common")]
983
9710
    fn compute_rings(&mut self) {
984
9710
        let params = HsDirParams::compute(&self.netdir.consensus, &self.netdir.params)
985
9710
            .expect("Invalid consensus");
986
        // TODO: see TODO by similar expect in new()
987

            
988
9710
        self.netdir.hsdir_rings =
989
10306
            Arc::new(params.map(|params| {
990
10086
                HsDirRing::compute(params, &self.netdir, self.prev_netdir.as_deref())
991
10086
            }));
992
9710
    }
993

            
994
    /// Return true if this are enough information in this directory
995
    /// to build multihop paths.
996
4
    pub fn have_enough_paths(&self) -> bool {
997
4
        self.netdir.have_enough_paths()
998
4
    }
999
    /// If this directory has enough information to build multihop
    /// circuits, return it.
9902
    pub fn unwrap_if_sufficient(
9902
        #[allow(unused_mut)] mut self,
9902
    ) -> std::result::Result<NetDir, PartialNetDir> {
9902
        if self.netdir.have_enough_paths() {
            #[cfg(feature = "hs-common")]
9710
            self.compute_rings();
9710
            Ok(self.netdir)
        } else {
192
            Err(self)
        }
9902
    }
}
impl MdReceiver for PartialNetDir {
102
    fn missing_microdescs(&self) -> Box<dyn Iterator<Item = &MdDigest> + '_> {
102
        self.netdir.missing_microdescs()
102
    }
386742
    fn add_microdesc(&mut self, md: Microdesc) -> bool {
386742
        self.netdir.add_microdesc(md)
386742
    }
282
    fn n_missing(&self) -> usize {
282
        self.netdir.n_missing()
282
    }
}
impl NetDir {
    /// Return the declared lifetime of this NetDir.
331
    pub fn lifetime(&self) -> &netstatus::Lifetime {
331
        self.consensus.lifetime()
331
    }
    /// Add `md` to this NetDir.
    ///
    /// Return true if we wanted it, and false otherwise.
386818
    fn add_arc_microdesc(&mut self, md: Arc<Microdesc>) -> bool {
386818
        if let Some(rsidx) = self.rsidx_by_missing.remove(md.digest()) {
370415
            assert_eq!(self.c_relays()[rsidx].md_digest(), md.digest());
            // There should never be two approved MDs in the same
            // consensus listing the same ID... but if there is,
            // we'll let the most recent one win.
370415
            self.rsidx_by_ed.insert(*md.ed25519_id(), rsidx);
            // Happy path: we did indeed want this one.
370415
            self.mds[rsidx] = Some(md);
            // Save some space in the missing-descriptor list.
370415
            if self.rsidx_by_missing.len() < self.rsidx_by_missing.capacity() / 4 {
18629
                self.rsidx_by_missing.shrink_to_fit();
351786
            }
370415
            return true;
16403
        }
        // Either we already had it, or we never wanted it at all.
16403
        false
386818
    }
    /// Construct a (possibly invalid) Relay object from a routerstatus and its
    /// index within the consensus.
50868366
    fn relay_from_rs_and_rsidx<'a>(
50868366
        &'a self,
50868366
        rs: &'a netstatus::MdRouterStatus,
50868366
        rsidx: RouterStatusIdx,
50868366
    ) -> UncheckedRelay<'a> {
50868366
        debug_assert_eq!(self.c_relays()[rsidx].rsa_identity(), rs.rsa_identity());
50868366
        let md = self.mds[rsidx].as_deref();
50868366
        if let Some(md) = md {
50864549
            debug_assert_eq!(rs.md_digest(), md.digest());
3817
        }
50868366
        UncheckedRelay {
50868366
            rs,
50868366
            md,
50868366
            #[cfg(feature = "geoip")]
50868366
            cc: self.country_codes.get(rsidx.0).copied().flatten(),
50868366
        }
50868366
    }
    /// Return the value of the hsdir_n_replicas param.
    #[cfg(feature = "hs-common")]
423
    fn n_replicas(&self) -> u8 {
423
        self.params
423
            .hsdir_n_replicas
423
            .get()
423
            .try_into()
423
            .expect("BoundedInt did not enforce bounds")
423
    }
    /// Return the spread parameter for the specified `op`.
    #[cfg(feature = "hs-common")]
423
    fn spread(&self, op: HsDirOp) -> usize {
423
        let spread = match op {
47
            HsDirOp::Download => self.params.hsdir_spread_fetch,
            #[cfg(feature = "hs-service")]
376
            HsDirOp::Upload => self.params.hsdir_spread_store,
        };
423
        spread
423
            .get()
423
            .try_into()
423
            .expect("BoundedInt did not enforce bounds!")
423
    }
    /// Select `spread` hsdir relays for the specified `hsid` from a given `ring`.
    ///
    /// Algorithm:
    ///
    /// for idx in 1..=n_replicas:
    ///       - let H = hsdir_ring::onion_service_index(id, replica, rand,
    ///         period).
    ///       - Find the position of H within hsdir_ring.
    ///       - Take elements from hsdir_ring starting at that position,
    ///         adding them to Dirs until we have added `spread` new elements
    ///         that were not there before.
    #[cfg(feature = "hs-common")]
423
    fn select_hsdirs<'h, 'r: 'h>(
423
        &'r self,
423
        hsid: HsBlindId,
423
        ring: &'h HsDirRing,
423
        spread: usize,
423
    ) -> impl Iterator<Item = Relay<'r>> + 'h {
423
        let n_replicas = self.n_replicas();
423
        (1..=n_replicas) // 1-indexed !
423
            .flat_map({
423
                let mut selected_nodes = HashSet::new();
36
                move |replica: u8| {
36
                    let hsdir_idx = hsdir_ring::service_hsdir_index(&hsid, replica, ring.params());
206
                    ring.ring_items_at(hsdir_idx, spread, |(hsdir_idx, _)| {
                        // According to rend-spec 2.2.3:
                        //                                                  ... If any of those
                        // nodes have already been selected for a lower-numbered replica of the
                        // service, any nodes already chosen are disregarded (i.e. skipped over)
                        // when choosing a replica's hsdir_spread_store nodes.
206
                        selected_nodes.insert(*hsdir_idx)
206
                    })
36
                    .collect::<Vec<_>>()
36
                }
            })
548
            .filter_map(move |(_hsdir_idx, rs_idx)| {
                // This ought not to be None but let's not panic or bail if it is
140
                self.relay_by_rs_idx(*rs_idx)
140
            })
423
    }
    /// Replace the overridden parameters in this netdir with `new_replacement`.
    ///
    /// After this function is done, the netdir's parameters will be those in
    /// the consensus, overridden by settings from `new_replacement`.  Any
    /// settings in the old replacement parameters will be discarded.
    pub fn replace_overridden_parameters(&mut self, new_replacement: &netstatus::NetParams<i32>) {
        // TODO(nickm): This is largely duplicate code from PartialNetDir::new().
        let mut new_params = NetParameters::default();
        let _ = new_params.saturating_update(self.consensus.params().iter());
        for u in new_params.saturating_update(new_replacement.iter()) {
            warn!("Unrecognized option: override_net_params.{}", u);
        }
        self.params = new_params;
    }
    /// Return an iterator over all Relay objects, including invalid ones
    /// that we can't use.
1283593
    pub fn all_relays(&self) -> impl Iterator<Item = UncheckedRelay<'_>> {
        // TODO: I'd like if we could memoize this so we don't have to
        // do so many hashtable lookups.
1283593
        self.c_relays()
1283593
            .iter_enumerated()
49642115
            .map(move |(rsidx, rs)| self.relay_from_rs_and_rsidx(rs, rsidx))
1283593
    }
    /// Return an iterator over all [usable](NetDir#usable) Relays.
1216736
    pub fn relays(&self) -> impl Iterator<Item = Relay<'_>> {
1216736
        self.all_relays().filter_map(UncheckedRelay::into_relay)
1216736
    }
    /// Look up a relay's [`Microdesc`] by its [`RouterStatusIdx`]
    #[cfg_attr(not(feature = "hs-common"), allow(dead_code))]
    pub(crate) fn md_by_rsidx(&self, rsidx: RouterStatusIdx) -> Option<&Microdesc> {
        self.mds.get(rsidx)?.as_deref()
    }
    /// Return a relay matching a given identity, if we have a
    /// _usable_ relay with that key.
    ///
    /// (Does not return [unusable](NetDir#usable) relays.)
    ///
    ///
    /// Note that a `None` answer is not always permanent: if a microdescriptor
    /// is subsequently added for a relay with this ID, the ID may become usable
    /// even if it was not usable before.
603909
    pub fn by_id<'a, T>(&self, id: T) -> Option<Relay<'_>>
603909
    where
603909
        T: Into<RelayIdRef<'a>>,
    {
603909
        let id = id.into();
603909
        let answer = match id {
603748
            RelayIdRef::Ed25519(ed25519) => {
603748
                let rsidx = *self.rsidx_by_ed.get(ed25519)?;
598662
                let rs = self.c_relays().get(rsidx).expect("Corrupt index");
598662
                self.relay_from_rs_and_rsidx(rs, rsidx).into_relay()?
            }
161
            RelayIdRef::Rsa(rsa) => self
161
                .by_rsa_id_unchecked(rsa)
161
                .and_then(UncheckedRelay::into_relay)?,
            other_type => self.relays().find(|r| r.has_identity(other_type))?,
        };
598676
        assert!(answer.has_identity(id));
598676
        Some(answer)
603909
    }
    /// Obtain a `Relay` given a `RouterStatusIdx`
    ///
    /// Differs from `relay_from_rs_and_rsi` as follows:
    ///  * That function expects the caller to already have an `MdRouterStatus`;
    ///    it checks with `debug_assert` that the relay in the netdir matches.
    ///  * That function panics if the `RouterStatusIdx` is invalid; this one returns `None`.
    ///  * That function returns an `UncheckedRelay`; this one a `Relay`.
    ///
    /// `None` could be returned here, even with a valid `rsi`,
    /// if `rsi` refers to an [unusable](NetDir#usable) relay.
    #[cfg_attr(not(feature = "hs-common"), allow(dead_code))]
3318
    pub(crate) fn relay_by_rs_idx(&self, rs_idx: RouterStatusIdx) -> Option<Relay<'_>> {
3318
        let rs = self.c_relays().get(rs_idx)?;
3318
        let md = self.mds.get(rs_idx)?.as_deref();
3318
        UncheckedRelay {
3318
            rs,
3318
            md,
3318
            #[cfg(feature = "geoip")]
3318
            cc: self.country_codes.get(rs_idx.0).copied().flatten(),
3318
        }
3318
        .into_relay()
3318
    }
    /// Return a relay with the same identities as those in `target`, if one
    /// exists.
    ///
    /// Does not return [unusable](NetDir#usable) relays.
    ///
    /// Note that a negative result from this method is not necessarily permanent:
    /// it may be the case that a relay exists,
    /// but we don't yet have enough information about it to know all of its IDs.
    /// To test whether a relay is *definitely* absent,
    /// use [`by_ids_detailed`](Self::by_ids_detailed)
    /// or [`ids_listed`](Self::ids_listed).
    ///
    /// # Limitations
    ///
    /// This will be very slow if `target` does not have an Ed25519 or RSA
    /// identity.
545900
    pub fn by_ids<T>(&self, target: &T) -> Option<Relay<'_>>
545900
    where
545900
        T: HasRelayIds + ?Sized,
    {
545900
        let mut identities = target.identities();
        // Don't try if there are no identities.
545900
        let first_id = identities.next()?;
        // Since there is at most one relay with each given ID type,
        // we only need to check the first relay we find.
545900
        let candidate = self.by_id(first_id)?;
545688
        if identities.all(|wanted_id| candidate.has_identity(wanted_id)) {
545686
            Some(candidate)
        } else {
2
            None
        }
545900
    }
    /// Check whether there is a relay that has at least one identity from
    /// `target`, and which _could_ have every identity from `target`.
    /// If so, return such a relay.
    ///
    /// Return `Ok(None)` if we did not find a relay with any identity from `target`.
    ///
    /// Return `RelayLookupError::Impossible` if we found a relay with at least
    /// one identity from `target`, but that relay's other identities contradict
    /// what we learned from `target`.
    ///
    /// Does not return [unusable](NetDir#usable) relays.
    ///
    /// (This function is only useful if you need to distinguish the
    /// "impossible" case from the "no such relay known" case.)
    ///
    /// # Limitations
    ///
    /// This will be very slow if `target` does not have an Ed25519 or RSA
    /// identity.
    //
    // TODO HS: This function could use a better name.
    //
    // TODO: We could remove the feature restriction here once we think this API is
    // stable.
    #[cfg(feature = "hs-common")]
155
    pub fn by_ids_detailed<T>(
155
        &self,
155
        target: &T,
155
    ) -> std::result::Result<Option<Relay<'_>>, RelayLookupError>
155
    where
155
        T: HasRelayIds + ?Sized,
    {
155
        let candidate = target
155
            .identities()
            // Find all the relays that share any identity with this set of identities.
306
            .filter_map(|id| self.by_id(id))
            // We might find the same relay more than once under a different
            // identity, so we remove the duplicates.
            //
            // Since there is at most one relay per rsa identity per consensus,
            // this is a true uniqueness check under current construction rules.
155
            .unique_by(|r| r.rs.rsa_identity())
            // If we find two or more distinct relays, then have a contradiction.
155
            .at_most_one()
155
            .map_err(|_| RelayLookupError::Impossible)?;
        // If we have no candidate, return None early.
155
        let candidate = match candidate {
10
            Some(relay) => relay,
145
            None => return Ok(None),
        };
        // Now we know we have a single candidate.  Make sure that it does not have any
        // identity that does not match the target.
10
        if target
10
            .identities()
14
            .all(|wanted_id| match candidate.identity(wanted_id.id_type()) {
                None => true,
14
                Some(id) => id == wanted_id,
14
            })
        {
6
            Ok(Some(candidate))
        } else {
4
            Err(RelayLookupError::Impossible)
        }
155
    }
    /// Return a boolean if this consensus definitely has (or does not have) a
    /// relay matching the listed identities.
    ///
    /// `Some(true)` indicates that the relay exists.
    /// `Some(false)` indicates that the relay definitely does not exist.
    /// `None` indicates that we can't yet tell whether such a relay exists,
    ///  due to missing information.
24493
    fn id_pair_listed(&self, ed_id: &Ed25519Identity, rsa_id: &RsaIdentity) -> Option<bool> {
24493
        let r = self.by_rsa_id_unchecked(rsa_id);
24493
        match r {
24162
            Some(unchecked) => {
24162
                if !unchecked.rs.ed25519_id_is_usable() {
                    return Some(false);
24162
                }
                // If md is present, then it's listed iff we have the right
                // ed id.  Otherwise we don't know if it's listed.
24674
                unchecked.md.map(|md| md.ed25519_id() == ed_id)
            }
            None => {
                // Definitely not listed.
331
                Some(false)
            }
        }
24493
    }
    /// Check whether a relay exists (or may exist)
    /// with the same identities as those in `target`.
    ///
    /// `Some(true)` indicates that the relay exists.
    /// `Some(false)` indicates that the relay definitely does not exist.
    /// `None` indicates that we can't yet tell whether such a relay exists,
    ///  due to missing information.
8434
    pub fn ids_listed<T>(&self, target: &T) -> Option<bool>
8434
    where
8434
        T: HasRelayIds + ?Sized,
    {
8434
        let rsa_id = target.rsa_identity();
8434
        let ed25519_id = target.ed_identity();
        // TODO: If we later support more identity key types, this will
        // become incorrect.  This assertion might help us recognize that case.
        const _: () = assert!(RelayIdType::COUNT == 2);
8434
        match (rsa_id, ed25519_id) {
8434
            (Some(r), Some(e)) => self.id_pair_listed(e, r),
            (Some(r), None) => Some(self.rsa_id_is_listed(r)),
            (None, Some(e)) => {
                if self.rsidx_by_ed.contains_key(e) {
                    Some(true)
                } else {
                    None
                }
            }
            (None, None) => None,
        }
8434
    }
    /// Return a (possibly [unusable](NetDir#usable)) relay with a given RSA identity.
    ///
    /// This API can be used to find information about a relay that is listed in
    /// the current consensus, even if we don't yet have enough information
    /// (like a microdescriptor) about the relay to use it.
275605
    #[cfg_attr(feature = "experimental-api", visibility::make(pub))]
275605
    #[cfg_attr(docsrs, doc(cfg(feature = "experimental-api")))]
275605
    fn by_rsa_id_unchecked(&self, rsa_id: &RsaIdentity) -> Option<UncheckedRelay<'_>> {
275605
        let rsidx = *self.rsidx_by_rsa.get(rsa_id)?;
275123
        let rs = self.c_relays().get(rsidx).expect("Corrupt index");
275123
        assert_eq!(rs.rsa_identity(), rsa_id);
275123
        Some(self.relay_from_rs_and_rsidx(rs, rsidx))
275605
    }
    /// Return the relay with a given RSA identity, if we have one
    /// and it is [usable](NetDir#usable).
12
    fn by_rsa_id(&self, rsa_id: &RsaIdentity) -> Option<Relay<'_>> {
12
        self.by_rsa_id_unchecked(rsa_id)?.into_relay()
12
    }
    /// Return true if `rsa_id` is listed in this directory, even if it isn't
    /// currently usable.
    ///
    /// (An "[unusable](NetDir#usable)" relay in this context is one for which we don't have full
    /// directory information.)
6
    #[cfg_attr(feature = "experimental-api", visibility::make(pub))]
6
    #[cfg_attr(docsrs, doc(cfg(feature = "experimental-api")))]
6
    fn rsa_id_is_listed(&self, rsa_id: &RsaIdentity) -> bool {
6
        self.by_rsa_id_unchecked(rsa_id).is_some()
6
    }
    /// List the hsdirs in this NetDir, that should be in the HSDir rings
    ///
    /// The results are not returned in any particular order.
    #[cfg(feature = "hs-common")]
10086
    fn all_hsdirs(&self) -> impl Iterator<Item = (RouterStatusIdx, Relay<'_>)> {
385573
        self.c_relays().iter_enumerated().filter_map(|(rsidx, rs)| {
385345
            let relay = self.relay_from_rs_and_rsidx(rs, rsidx);
385345
            relay.is_hsdir_for_ring().then_some(())?;
94374
            let relay = relay.into_relay()?;
94366
            Some((rsidx, relay))
385345
        })
10086
    }
    /// Return the parameters from the consensus, clamped to the
    /// correct ranges, with defaults filled in.
    ///
    /// NOTE: that unsupported parameters aren't returned here; only those
    /// values configured in the `params` module are available.
606007
    pub fn params(&self) -> &NetParameters {
606007
        &self.params
606007
    }
    /// Return a [`ProtoStatus`](netstatus::ProtoStatus) that lists the
    /// network's current requirements and recommendations for the list of
    /// protocols that every relay must implement.
    //
    // TODO HS: I am not sure this is the right API; other alternatives would be:
    //    * To expose the _required_ relay protocol list instead (since that's all that
    //      onion service implementations need).
    //    * To expose the client protocol list as well (for symmetry).
    //    * To expose the MdConsensus instead (since that's more general, although
    //      it restricts the future evolution of this API).
    //
    // I think that this is a reasonably good compromise for now, but I'm going
    // to put it behind the `hs-common` feature to give us time to consider more.
    #[cfg(feature = "hs-common")]
141
    pub fn relay_protocol_status(&self) -> &netstatus::ProtoStatus {
141
        self.consensus.relay_protocol_status()
141
    }
    /// Return a [`ProtoStatus`](netstatus::ProtoStatus) that lists the
    /// network's current requirements and recommendations for the list of
    /// protocols that every relay must implement.
    //
    // TODO HS: See notes on relay_protocol_status above.
    #[cfg(feature = "hs-common")]
    pub fn client_protocol_status(&self) -> &netstatus::ProtoStatus {
        self.consensus.client_protocol_status()
    }
    /// Construct a `CircTarget` from an externally provided list of link specifiers,
    /// and an externally provided onion key.
    ///
    /// This method is used in the onion service protocol,
    /// where introduction points and rendezvous points are specified using these inputs.
    ///
    /// This function is a member of `NetDir` so that it can provide a reasonable list of
    /// [`Protocols`](tor_protover::Protocols) capabilities for the generated `CircTarget`.
    /// It does not (and should not!) look up anything else from the directory.
    #[cfg(feature = "hs-common")]
141
    pub fn circ_target_from_verbatim_linkspecs(
141
        &self,
141
        linkspecs: &[tor_linkspec::EncodedLinkSpec],
141
        ntor_onion_key: &curve25519::PublicKey,
141
    ) -> StdResult<VerbatimLinkSpecCircTarget<OwnedCircTarget>, VerbatimCircTargetDecodeError> {
        use VerbatimCircTargetDecodeError as E;
        use tor_linkspec::CircTarget as _;
        use tor_linkspec::decode::Strictness;
141
        let mut bld = OwnedCircTarget::builder();
        use tor_error::into_internal;
141
        *bld.chan_target() =
141
            OwnedChanTargetBuilder::from_encoded_linkspecs(Strictness::Standard, linkspecs)?;
141
        let protocols = {
141
            let chan_target = bld.chan_target().build().map_err(into_internal!(
                "from_encoded_linkspecs gave an invalid output"
            ))?;
141
            match self
141
                .by_ids_detailed(&chan_target)
141
                .map_err(E::ImpossibleIds)?
            {
                Some(relay) => relay.protovers().clone(),
141
                None => self.relay_protocol_status().required_protocols().clone(),
            }
        };
141
        bld.protocols(protocols);
141
        bld.ntor_onion_key(*ntor_onion_key);
141
        Ok(VerbatimLinkSpecCircTarget::new(
141
            bld.build()
141
                .map_err(into_internal!("Failed to construct a valid circtarget"))?,
141
            linkspecs.to_vec(),
        ))
141
    }
    /// Return weighted the fraction of relays we can use.  We only
    /// consider relays that match the predicate `usable`.  We weight
    /// this bandwidth according to the provided `role`.
    ///
    /// If _no_ matching relays in the consensus have a nonzero
    /// weighted bandwidth value, we fall back to looking at the
    /// unweighted fraction of matching relays.
    ///
    /// If there are no matching relays in the consensus, we return 0.0.
29119
    fn frac_for_role<'a, F>(&'a self, role: WeightRole, usable: F) -> f64
29119
    where
29119
        F: Fn(&UncheckedRelay<'a>) -> bool,
    {
29119
        let mut total_weight = 0_u64;
29119
        let mut have_weight = 0_u64;
29119
        let mut have_count = 0_usize;
29119
        let mut total_count = 0_usize;
1106574
        for r in self.all_relays() {
1106574
            if !usable(&r) {
366790
                continue;
739784
            }
739784
            let w = self.weights.weight_rs_for_role(r.rs, role);
739784
            total_weight += w;
739784
            total_count += 1;
739784
            if r.is_usable() {
736912
                have_weight += w;
736912
                have_count += 1;
736912
            }
        }
29119
        if total_weight > 0 {
            // The consensus lists some weighted bandwidth so return the
            // fraction of the weighted bandwidth for which we have
            // descriptors.
29119
            (have_weight as f64) / (total_weight as f64)
        } else if total_count > 0 {
            // The consensus lists no weighted bandwidth for these relays,
            // but at least it does list relays. Return the fraction of
            // relays for which it we have descriptors.
            (have_count as f64) / (total_count as f64)
        } else {
            // There are no relays of this kind in the consensus.  Return
            // 0.0, to avoid dividing by zero and giving NaN.
            0.0
        }
29119
    }
    /// Return the estimated fraction of possible paths that we have
    /// enough microdescriptors to build.
9908
    fn frac_usable_paths(&self) -> f64 {
        // TODO #504, TODO SPEC: We may want to add a set of is_flagged_fast() and/or
        // is_flagged_stable() checks here.  This will require spec clarification.
371780
        let f_g = self.frac_for_role(WeightRole::Guard, |u| {
371551
            u.low_level_details().is_suitable_as_guard()
371551
        });
9908
        let f_m = self.frac_for_role(WeightRole::Middle, |_| true);
108371
        let f_e = if self.all_relays().any(|u| u.rs.is_flagged_exit()) {
363448
            self.frac_for_role(WeightRole::Exit, |u| u.rs.is_flagged_exit())
        } else {
            // If there are no exits at all, we use f_m here.
611
            f_m
        };
9908
        f_g * f_m * f_e
9908
    }
    /// Return true if there is enough information in this NetDir to build
    /// multihop circuits.
9906
    fn have_enough_paths(&self) -> bool {
        // TODO-A001: This should check for our guards as well, and
        // make sure that if they're listed in the consensus, we have
        // the descriptors for them.
        // If we can build a randomly chosen path with at least this
        // probability, we know enough information to participate
        // on the network.
9906
        let min_frac_paths: f64 = self.params().min_circuit_path_threshold.as_fraction();
        // What fraction of paths can we build?
9906
        let available = self.frac_usable_paths();
9906
        available >= min_frac_paths
9906
    }
    /// Choose a relay at random.
    ///
    /// Each relay is chosen with probability proportional to its weight
    /// in the role `role`, and is only selected if the predicate `usable`
    /// returns true for it.
    ///
    /// This function returns None if (and only if) there are no relays
    /// with nonzero weight where `usable` returned true.
    //
    // TODO this API, with the `usable` closure, invites mistakes where we fail to
    // check conditions that are implied by the role we have selected for the relay:
    // call sites must include a call to `Relay::is_polarity_inverter()` or whatever.
    // IMO the `WeightRole` ought to imply a condition (and it should therefore probably
    // be renamed.)  -Diziet
59604
    pub fn pick_relay<'a, R, P>(
59604
        &'a self,
59604
        rng: &mut R,
59604
        role: WeightRole,
59604
        usable: P,
59604
    ) -> Option<Relay<'a>>
59604
    where
59604
        R: rand::Rng,
59604
        P: FnMut(&Relay<'a>) -> bool,
    {
59604
        let relays: Vec<_> = self.relays().filter(usable).collect();
        // This algorithm uses rand::distr::WeightedIndex, and uses
        // gives O(n) time and space  to build the index, plus O(log n)
        // sampling time.
        //
        // We might be better off building a WeightedIndex in advance
        // for each `role`, and then sampling it repeatedly until we
        // get a relay that satisfies `usable`.  Or we might not --
        // that depends heavily on the actual particulars of our
        // inputs.  We probably shouldn't make any changes there
        // unless profiling tells us that this function is in a hot
        // path.
        //
        // The C Tor sampling implementation goes through some trouble
        // here to try to make its path selection constant-time.  I
        // believe that there is no actual remotely exploitable
        // side-channel here however.  It could be worth analyzing in
        // the future.
        //
        // This code will give the wrong result if the total of all weights
        // can exceed u64::MAX.  We make sure that can't happen when we
        // set up `self.weights`.
1052360
        match relays[..].choose_weighted(rng, |r| self.weights.weight_rs_for_role(r.rs, role)) {
59538
            Ok(relay) => Some(relay.clone()),
            Err(WeightError::InsufficientNonZero) => {
                if relays.is_empty() {
                    None
                } else {
                    warn!(?self.weights, ?role,
                          "After filtering, all {} relays had zero weight. Choosing one at random. See bug #1907.",
                          relays.len());
                    relays.choose(rng).cloned()
                }
            }
66
            Err(e) => {
66
                warn_report!(e, "Unexpected error while sampling a relay");
66
                None
            }
        }
59604
    }
    /// Choose `n` relay at random.
    ///
    /// Each relay is chosen with probability proportional to its weight
    /// in the role `role`, and is only selected if the predicate `usable`
    /// returns true for it.
    ///
    /// Relays are chosen without replacement: no relay will be
    /// returned twice. Therefore, the resulting vector may be smaller
    /// than `n` if we happen to have fewer than `n` appropriate relays.
    ///
    /// This function returns an empty vector if (and only if) there
    /// are no relays with nonzero weight where `usable` returned
    /// true.
    #[allow(clippy::cognitive_complexity)] // all due to tracing crate.
29096
    pub fn pick_n_relays<'a, R, P>(
29096
        &'a self,
29096
        rng: &mut R,
29096
        n: usize,
29096
        role: WeightRole,
29096
        usable: P,
29096
    ) -> Vec<Relay<'a>>
29096
    where
29096
        R: rand::Rng,
29096
        P: FnMut(&Relay<'a>) -> bool,
    {
29096
        let relays: Vec<_> = self.relays().filter(usable).collect();
        // NOTE: See discussion in pick_relay().
206162
        let mut relays = match relays[..].choose_multiple_weighted(rng, n, |r| {
206162
            self.weights.weight_rs_for_role(r.rs, role) as f64
206162
        }) {
            Err(WeightError::InsufficientNonZero) => {
                // Too few relays had nonzero weights: return all of those that are okay.
                // (This is behavior used to come up with rand 0.9; it no longer does.
                // We still detect it.)
                let remaining: Vec<_> = relays
                    .iter()
                    .filter(|r| self.weights.weight_rs_for_role(r.rs, role) > 0)
                    .cloned()
                    .collect();
                if remaining.is_empty() {
                    warn!(?self.weights, ?role,
                          "After filtering, all {} relays had zero weight! Picking some at random. See bug #1907.",
                          relays.len());
                    if relays.len() >= n {
                        relays.choose_multiple(rng, n).cloned().collect()
                    } else {
                        relays
                    }
                } else {
                    warn!(?self.weights, ?role,
                          "After filtering, only had {}/{} relays with nonzero weight. Returning them all. See bug #1907.",
                           remaining.len(), relays.len());
                    remaining
                }
            }
            Err(e) => {
                warn_report!(e, "Unexpected error while sampling a set of relays");
                Vec::new()
            }
29096
            Ok(iter) => {
29096
                let selection: Vec<_> = iter.map(Relay::clone).collect();
29096
                if selection.len() < n && selection.len() < relays.len() {
                    warn!(?self.weights, ?role,
                          "choose_multiple_weighted returned only {returned}, despite requesting {n}, \
                          and having {filtered_len} available after filtering. See bug #1907.",
                          returned=selection.len(), filtered_len=relays.len());
29096
                }
29096
                selection
            }
        };
29096
        relays.shuffle(rng);
29096
        relays
29096
    }
    /// Compute the weight with which `relay` will be selected for a given
    /// `role`.
35252
    pub fn relay_weight<'a>(&'a self, relay: &Relay<'a>, role: WeightRole) -> RelayWeight {
35252
        RelayWeight(self.weights.weight_rs_for_role(relay.rs, role))
35252
    }
    /// Compute the total weight with which any relay matching `usable`
    /// will be selected for a given `role`.
    ///
    /// Note: because this function is used to assess the total
    /// properties of the consensus, the `usable` predicate takes a
    /// [`MdRouterStatus`] rather than a [`Relay`].
23632
    pub fn total_weight<P>(&self, role: WeightRole, usable: P) -> RelayWeight
23632
    where
23632
        P: Fn(&UncheckedRelay<'_>) -> bool,
    {
23632
        self.all_relays()
529128
            .filter_map(|unchecked| {
529128
                if usable(&unchecked) {
226964
                    Some(RelayWeight(
226964
                        self.weights.weight_rs_for_role(unchecked.rs, role),
226964
                    ))
                } else {
302164
                    None
                }
529128
            })
23632
            .sum()
23632
    }
    /// Compute the weight with which a relay with ID `rsa_id` would be
    /// selected for a given `role`.
    ///
    /// Note that weight returned by this function assumes that the
    /// relay with that ID is actually [usable](NetDir#usable); if it isn't usable,
    /// then other weight-related functions will call its weight zero.
250749
    pub fn weight_by_rsa_id(&self, rsa_id: &RsaIdentity, role: WeightRole) -> Option<RelayWeight> {
250749
        self.by_rsa_id_unchecked(rsa_id)
256085
            .map(|unchecked| RelayWeight(self.weights.weight_rs_for_role(unchecked.rs, role)))
250749
    }
    /// Return all relays in this NetDir known to be in the same family as
    /// `relay`.
    ///
    /// This list of members will **not** necessarily include `relay` itself.
    ///
    /// # Limitations
    ///
    /// Two relays only belong to the same family if _each_ relay
    /// claims to share a family with the other.  But if we are
    /// missing a microdescriptor for one of the relays listed by this
    /// relay, we cannot know whether it acknowledges family
    /// membership with this relay or not.  Therefore, this function
    /// can omit family members for which there is not (as yet) any
    /// Relay object.
4
    pub fn known_family_members<'a>(
4
        &'a self,
4
        relay: &'a Relay<'a>,
4
    ) -> impl Iterator<Item = Relay<'a>> {
4
        let relay_rsa_id = relay.rsa_id();
10
        relay.md.family().members().filter_map(move |other_rsa_id| {
8
            self.by_rsa_id(other_rsa_id)
8
                .filter(|other_relay| other_relay.md.family().contains(relay_rsa_id))
8
        })
4
    }
    /// Return the current hidden service directory "time period".
    ///
    /// Specifically, this returns the time period that contains the beginning
    /// of the validity period of this `NetDir`'s consensus.  That time period
    /// is the one we use when acting as an hidden service client.
    #[cfg(feature = "hs-common")]
1410
    pub fn hs_time_period(&self) -> TimePeriod {
1410
        self.hsdir_rings.current.time_period()
1410
    }
    /// Return the [`HsDirParams`] of all the relevant hidden service directory "time periods"
    ///
    /// This includes the current time period (as from
    /// [`.hs_time_period`](NetDir::hs_time_period))
    /// plus additional time periods that we publish descriptors for when we are
    /// acting as a hidden service.
    #[cfg(feature = "hs-service")]
376
    pub fn hs_all_time_periods(&self) -> Vec<HsDirParams> {
376
        self.hsdir_rings
376
            .iter()
760
            .map(|r| r.params().clone())
376
            .collect()
376
    }
    /// Return the relays in this network directory that will be used as hidden service directories
    ///
    /// These are suitable to retrieve a given onion service's descriptor at a given time period.
    #[cfg(feature = "hs-common")]
2
    pub fn hs_dirs_download<'r, R>(
2
        &'r self,
2
        hsid: HsBlindId,
2
        period: TimePeriod,
2
        rng: &mut R,
2
    ) -> std::result::Result<Vec<Relay<'r>>, Bug>
2
    where
2
        R: rand::Rng,
    {
        // Algorithm:
        //
        // 1. Determine which HsDirRing to use, based on the time period.
        // 2. Find the shared random value that's associated with that HsDirRing.
        // 3. Choose spread = the parameter `hsdir_spread_fetch`
        // 4. Let n_replicas = the parameter `hsdir_n_replicas`.
        // 5. Initialize Dirs = []
        // 6. for idx in 1..=n_replicas:
        //       - let H = hsdir_ring::onion_service_index(id, replica, rand,
        //         period).
        //       - Find the position of H within hsdir_ring.
        //       - Take elements from hsdir_ring starting at that position,
        //         adding them to Dirs until we have added `spread` new elements
        //         that were not there before.
        // 7. Shuffle Dirs
        // 8. return Dirs.
2
        let spread = self.spread(HsDirOp::Download);
        // When downloading, only look at relays on current ring.
2
        let ring = &self.hsdir_rings.current;
2
        if ring.params().time_period != period {
            return Err(internal!(
                "our current ring is not associated with the requested time period!"
            ));
2
        }
2
        let mut hs_dirs = self.select_hsdirs(hsid, ring, spread).collect_vec();
        // When downloading, the order of the returned relays is random.
2
        hs_dirs.shuffle(rng);
2
        Ok(hs_dirs)
2
    }
    /// Return the relays in this network directory that will be used as hidden service directories
    ///
    /// Returns the relays that are suitable for storing a given onion service's descriptors at the
    /// given time period.
    #[cfg(feature = "hs-service")]
376
    pub fn hs_dirs_upload(
376
        &self,
376
        hsid: HsBlindId,
376
        period: TimePeriod,
376
    ) -> std::result::Result<impl Iterator<Item = Relay<'_>>, Bug> {
        // Algorithm:
        //
        // 1. Choose spread = the parameter `hsdir_spread_store`
        // 2. Determine which HsDirRing to use, based on the time period.
        // 3. Find the shared random value that's associated with that HsDirRing.
        // 4. Let n_replicas = the parameter `hsdir_n_replicas`.
        // 5. Initialize Dirs = []
        // 6. for idx in 1..=n_replicas:
        //       - let H = hsdir_ring::onion_service_index(id, replica, rand,
        //         period).
        //       - Find the position of H within hsdir_ring.
        //       - Take elements from hsdir_ring starting at that position,
        //         adding them to Dirs until we have added `spread` new elements
        //         that were not there before.
        // 3. return Dirs.
376
        let spread = self.spread(HsDirOp::Upload);
        // For each HsBlindId, determine which HsDirRing to use.
376
        let rings = self
376
            .hsdir_rings
376
            .iter()
384
            .filter_map(move |ring| {
                // Make sure the ring matches the TP of the hsid it's matched with.
376
                (ring.params().time_period == period).then_some((ring, hsid, period))
376
            })
376
            .collect::<Vec<_>>();
        // The specified period should have an associated ring.
384
        if !rings.iter().any(|(_, _, tp)| *tp == period) {
            return Err(internal!(
                "the specified time period does not have an associated ring"
            ));
376
        };
        // Now that we've matched each `hsid` with the ring associated with its TP, we can start
        // selecting replicas from each ring.
384
        Ok(rings.into_iter().flat_map(move |(ring, hsid, period)| {
16
            assert_eq!(period, ring.params().time_period());
16
            self.select_hsdirs(hsid, ring, spread)
16
        }))
376
    }
    /// Return the relays in this network directory that will be used as hidden service directories
    ///
    /// Depending on `op`,
    /// these are suitable to either store, or retrieve, a
    /// given onion service's descriptor at a given time period.
    ///
    /// When `op` is `Download`, the order is random.
    /// When `op` is `Upload`, the order is not specified.
    ///
    /// Return an error if the time period is not one returned by
    /// `onion_service_time_period` or `onion_service_secondary_time_periods`.
    //
    // TODO: make HsDirOp pub(crate) once this is removed
    #[cfg(feature = "hs-common")]
    #[deprecated(note = "Use hs_dirs_upload or hs_dirs_download instead")]
4
    pub fn hs_dirs<'r, R>(&'r self, hsid: &HsBlindId, op: HsDirOp, rng: &mut R) -> Vec<Relay<'r>>
4
    where
4
        R: rand::Rng,
    {
        // Algorithm:
        //
        // 1. Determine which HsDirRing to use, based on the time period.
        // 2. Find the shared random value that's associated with that HsDirRing.
        // 3. Choose spread = the parameter `hsdir_spread_store` or
        //    `hsdir_spread_fetch` based on `op`.
        // 4. Let n_replicas = the parameter `hsdir_n_replicas`.
        // 5. Initialize Dirs = []
        // 6. for idx in 1..=n_replicas:
        //       - let H = hsdir_ring::onion_service_index(id, replica, rand,
        //         period).
        //       - Find the position of H within hsdir_ring.
        //       - Take elements from hsdir_ring starting at that position,
        //         adding them to Dirs until we have added `spread` new elements
        //         that were not there before.
        // 7. return Dirs.
4
        let n_replicas = self
4
            .params
4
            .hsdir_n_replicas
4
            .get()
4
            .try_into()
4
            .expect("BoundedInt did not enforce bounds");
4
        let spread = match op {
2
            HsDirOp::Download => self.params.hsdir_spread_fetch,
            #[cfg(feature = "hs-service")]
2
            HsDirOp::Upload => self.params.hsdir_spread_store,
        };
4
        let spread = spread
4
            .get()
4
            .try_into()
4
            .expect("BoundedInt did not enforce bounds!");
        // TODO: I may be wrong here but I suspect that this function may
        // need refactoring so that it does not look at _all_ of the HsDirRings,
        // but only at the ones that corresponds to time periods for which
        // HsBlindId is valid.  Or I could be mistaken, in which case we should
        // have a comment to explain why I am, since the logic is subtle.
        // (For clients, there is only one ring.) -nickm
        //
        // (Actually, there is no need to follow through with the above TODO,
        // since this function is deprecated, and not used anywhere but the
        // tests.)
4
        let mut hs_dirs = self
4
            .hsdir_rings
4
            .iter_for_op(op)
4
            .cartesian_product(1..=n_replicas) // 1-indexed !
4
            .flat_map({
4
                let mut selected_nodes = HashSet::new();
8
                move |(ring, replica): (&HsDirRing, u8)| {
8
                    let hsdir_idx = hsdir_ring::service_hsdir_index(hsid, replica, ring.params());
40
                    ring.ring_items_at(hsdir_idx, spread, |(hsdir_idx, _)| {
                        // According to rend-spec 2.2.3:
                        //                                                  ... If any of those
                        // nodes have already been selected for a lower-numbered replica of the
                        // service, any nodes already chosen are disregarded (i.e. skipped over)
                        // when choosing a replica's hsdir_spread_store nodes.
40
                        selected_nodes.insert(*hsdir_idx)
40
                    })
8
                    .collect::<Vec<_>>()
8
                }
            })
28
            .filter_map(|(_hsdir_idx, rs_idx)| {
                // This ought not to be None but let's not panic or bail if it is
28
                self.relay_by_rs_idx(*rs_idx)
28
            })
4
            .collect_vec();
4
        match op {
2
            HsDirOp::Download => {
2
                // When `op` is `Download`, the order is random.
2
                hs_dirs.shuffle(rng);
2
            }
            #[cfg(feature = "hs-service")]
2
            HsDirOp::Upload => {
2
                // When `op` is `Upload`, the order is not specified.
2
            }
        }
4
        hs_dirs
4
    }
}
impl MdReceiver for NetDir {
151
    fn missing_microdescs(&self) -> Box<dyn Iterator<Item = &MdDigest> + '_> {
151
        Box::new(self.rsidx_by_missing.keys())
151
    }
386742
    fn add_microdesc(&mut self, md: Microdesc) -> bool {
386742
        self.add_arc_microdesc(Arc::new(md))
386742
    }
282
    fn n_missing(&self) -> usize {
282
        self.rsidx_by_missing.len()
282
    }
}
impl<'a> UncheckedRelay<'a> {
    /// Return an [`UncheckedRelayDetails`](details::UncheckedRelayDetails) for this relay.
    ///
    /// Callers should generally avoid using this information directly if they can;
    /// it's better to use a higher-level function that exposes semantic information
    /// rather than these properties.
1015545
    pub fn low_level_details(&self) -> details::UncheckedRelayDetails<'_> {
1015545
        details::UncheckedRelayDetails(self)
1015545
    }
    /// Return true if this relay is valid and [usable](NetDir#usable).
    ///
    /// This function should return `true` for every Relay we expose
    /// to the user.
49178313
    pub fn is_usable(&self) -> bool {
        // No need to check for 'valid' or 'running': they are implicit.
49178313
        self.md.is_some() && self.rs.ed25519_id_is_usable()
49178313
    }
    /// If this is [usable](NetDir#usable), return a corresponding Relay object.
48438529
    pub fn into_relay(self) -> Option<Relay<'a>> {
48438529
        if self.is_usable() {
            Some(Relay {
48438505
                rs: self.rs,
48438505
                md: self.md?,
                #[cfg(feature = "geoip")]
48438505
                cc: self.cc,
            })
        } else {
24
            None
        }
48438529
    }
    /// Return true if this relay is a hidden service directory
    ///
    /// Ie, if it is to be included in the hsdir ring.
    #[cfg(feature = "hs-common")]
385345
    pub(crate) fn is_hsdir_for_ring(&self) -> bool {
        // TODO are there any other flags should we check?
        // rend-spec-v3 2.2.3 says just
        //   "each node listed in the current consensus with the HSDir flag"
        // Do we need to check ed25519_id_is_usable ?
        // See also https://gitlab.torproject.org/tpo/core/arti/-/issues/504
385345
        self.rs.is_flagged_hsdir()
385345
    }
}
impl<'a> Relay<'a> {
    /// Return a [`RelayDetails`](details::RelayDetails) for this relay.
    ///
    /// Callers should generally avoid using this information directly if they can;
    /// it's better to use a higher-level function that exposes semantic information
    /// rather than these properties.
100265200
    pub fn low_level_details(&self) -> details::RelayDetails<'_> {
100265200
        details::RelayDetails(self)
100265200
    }
    /// Return the Ed25519 ID for this relay.
148004176
    pub fn id(&self) -> &Ed25519Identity {
148004176
        self.md.ed25519_id()
148004176
    }
    /// Return the RsaIdentity for this relay.
95791118
    pub fn rsa_id(&self) -> &RsaIdentity {
95791118
        self.rs.rsa_identity()
95791118
    }
    /// Return a reference to this relay's "router status" entry in
    /// the consensus.
    ///
    /// The router status entry contains information about the relay
    /// that the authorities voted on directly.  For most use cases,
    /// you shouldn't need them.
    ///
    /// This function is only available if the crate was built with
    /// its `experimental-api` feature.
    #[cfg(feature = "experimental-api")]
4
    pub fn rs(&self) -> &netstatus::MdRouterStatus {
4
        self.rs
4
    }
    /// Return a reference to this relay's "microdescriptor" entry in
    /// the consensus.
    ///
    /// A "microdescriptor" is a synopsis of the information about a relay,
    /// used to determine its capabilities and route traffic through it.
    /// For most use cases, you shouldn't need it.
    ///
    /// This function is only available if the crate was built with
    /// its `experimental-api` feature.
    #[cfg(feature = "experimental-api")]
4
    pub fn md(&self) -> &Microdesc {
4
        self.md
4
    }
}
/// An error value returned from [`NetDir::by_ids_detailed`].
#[cfg(feature = "hs-common")]
#[derive(Clone, Debug, thiserror::Error)]
#[non_exhaustive]
pub enum RelayLookupError {
    /// We found a relay whose presence indicates that the provided set of
    /// identities is impossible to resolve.
    #[error("Provided set of identities is impossible according to consensus.")]
    Impossible,
}
impl<'a> HasAddrs for Relay<'a> {
135565714
    fn addrs(&self) -> impl Iterator<Item = std::net::SocketAddr> {
135565714
        self.rs.addrs()
135565714
    }
}
#[cfg(feature = "geoip")]
impl<'a> HasCountryCode for Relay<'a> {
846
    fn country_code(&self) -> Option<CountryCode> {
846
        self.cc
846
    }
}
impl<'a> tor_linkspec::HasRelayIdsLegacy for Relay<'a> {
147992414
    fn ed_identity(&self) -> &Ed25519Identity {
147992414
        self.id()
147992414
    }
44777809
    fn rsa_identity(&self) -> &RsaIdentity {
44777809
        self.rsa_id()
44777809
    }
}
impl<'a> HasRelayIds for UncheckedRelay<'a> {
    fn identity(&self, key_type: RelayIdType) -> Option<RelayIdRef<'_>> {
        match key_type {
            RelayIdType::Ed25519 if self.rs.ed25519_id_is_usable() => {
                self.md.map(|m| m.ed25519_id().into())
            }
            RelayIdType::Rsa => Some(self.rs.rsa_identity().into()),
            _ => None,
        }
    }
}
#[cfg(feature = "geoip")]
impl<'a> HasCountryCode for UncheckedRelay<'a> {
    fn country_code(&self) -> Option<CountryCode> {
        self.cc
    }
}
impl<'a> DirectChanMethodsHelper for Relay<'a> {}
impl<'a> ChanTarget for Relay<'a> {}
impl<'a> tor_linkspec::CircTarget for Relay<'a> {
1760009
    fn ntor_onion_key(&self) -> &ll::pk::curve25519::PublicKey {
1760009
        self.md.ntor_key()
1760009
    }
1760009
    fn protovers(&self) -> &tor_protover::Protocols {
1760009
        self.rs.protovers()
1760009
    }
}
#[cfg(test)]
mod test {
    // @@ begin test lint list maintained by maint/add_warning @@
    #![allow(clippy::bool_assert_comparison)]
    #![allow(clippy::clone_on_copy)]
    #![allow(clippy::dbg_macro)]
    #![allow(clippy::mixed_attributes_style)]
    #![allow(clippy::print_stderr)]
    #![allow(clippy::print_stdout)]
    #![allow(clippy::single_char_pattern)]
    #![allow(clippy::unwrap_used)]
    #![allow(clippy::unchecked_time_subtraction)]
    #![allow(clippy::useless_vec)]
    #![allow(clippy::needless_pass_by_value)]
    //! <!-- @@ end test lint list maintained by maint/add_warning @@ -->
    #![allow(clippy::cognitive_complexity)]
    use super::*;
    use crate::testnet::*;
    use float_eq::assert_float_eq;
    use std::collections::HashSet;
    use std::time::Duration;
    use tor_basic_utils::test_rng::{self, testing_rng};
    use tor_linkspec::{RelayIdType, RelayIds};
    #[cfg(feature = "hs-common")]
    fn dummy_hs_blind_id() -> HsBlindId {
        let hsid = [2, 1, 1, 1].iter().cycle().take(32).cloned().collect_vec();
        let hsid = Ed25519Identity::new(hsid[..].try_into().unwrap());
        HsBlindId::from(hsid)
    }
    // Basic functionality for a partial netdir: Add microdescriptors,
    // then you have a netdir.
    #[test]
    fn partial_netdir() {
        let (consensus, microdescs) = construct_network().unwrap();
        let dir = PartialNetDir::new(consensus, None);
        // Check the lifetime
        let lifetime = dir.lifetime();
        assert_eq!(
            lifetime
                .valid_until()
                .duration_since(lifetime.valid_after())
                .unwrap(),
            Duration::new(86400, 0)
        );
        // No microdescriptors, so we don't have enough paths, and can't
        // advance.
        assert!(!dir.have_enough_paths());
        let mut dir = match dir.unwrap_if_sufficient() {
            Ok(_) => panic!(),
            Err(d) => d,
        };
        let missing: HashSet<_> = dir.missing_microdescs().collect();
        assert_eq!(missing.len(), 40);
        assert_eq!(missing.len(), dir.netdir.c_relays().len());
        for md in &microdescs {
            assert!(missing.contains(md.digest()));
        }
        // Now add all the mds and try again.
        for md in microdescs {
            let wanted = dir.add_microdesc(md);
            assert!(wanted);
        }
        let missing: HashSet<_> = dir.missing_microdescs().collect();
        assert!(missing.is_empty());
        assert!(dir.have_enough_paths());
        let _complete = match dir.unwrap_if_sufficient() {
            Ok(d) => d,
            Err(_) => panic!(),
        };
    }
    #[test]
    fn override_params() {
        let (consensus, _microdescs) = construct_network().unwrap();
        let override_p = "bwweightscale=2 doesnotexist=77 circwindow=500"
            .parse()
            .unwrap();
        let dir = PartialNetDir::new(consensus.clone(), Some(&override_p));
        let params = &dir.netdir.params;
        assert_eq!(params.bw_weight_scale.get(), 2);
        assert_eq!(params.circuit_window.get(), 500_i32);
        // try again without the override.
        let dir = PartialNetDir::new(consensus, None);
        let params = &dir.netdir.params;
        assert_eq!(params.bw_weight_scale.get(), 1_i32);
        assert_eq!(params.circuit_window.get(), 1000_i32);
    }
    #[test]
    fn fill_from_previous() {
        let (consensus, microdescs) = construct_network().unwrap();
        let mut dir = PartialNetDir::new(consensus.clone(), None);
        for md in microdescs.iter().skip(2) {
            let wanted = dir.add_microdesc(md.clone());
            assert!(wanted);
        }
        let dir1 = dir.unwrap_if_sufficient().unwrap();
        assert_eq!(dir1.missing_microdescs().count(), 2);
        let mut dir = PartialNetDir::new(consensus, None);
        assert_eq!(dir.missing_microdescs().count(), 40);
        dir.fill_from_previous_netdir(Arc::new(dir1));
        assert_eq!(dir.missing_microdescs().count(), 2);
    }
    #[test]
    fn path_count() {
        let low_threshold = "min_paths_for_circs_pct=64".parse().unwrap();
        let high_threshold = "min_paths_for_circs_pct=65".parse().unwrap();
        let (consensus, microdescs) = construct_network().unwrap();
        let mut dir = PartialNetDir::new(consensus.clone(), Some(&low_threshold));
        for (pos, md) in microdescs.iter().enumerate() {
            if pos % 7 == 2 {
                continue; // skip a few relays.
            }
            dir.add_microdesc(md.clone());
        }
        let dir = dir.unwrap_if_sufficient().unwrap();
        // We  have 40 relays that we know about from the consensus.
        assert_eq!(dir.all_relays().count(), 40);
        // But only 34 are usable.
        assert_eq!(dir.relays().count(), 34);
        // For guards: mds 20..=39 correspond to Guard relays.
        // Their bandwidth is 2*(1000+2000+...10000) = 110_000.
        // We skipped 23, 30, and 37.  They have bandwidth
        // 4000 + 1000 + 8000 = 13_000.  So our fractional bandwidth
        // should be (110-13)/110.
        let f = dir.frac_for_role(WeightRole::Guard, |u| u.rs.is_flagged_guard());
        assert!(((97.0 / 110.0) - f).abs() < 0.000001);
        // For exits: mds 10..=19 and 30..=39 correspond to Exit relays.
        // We skipped 16, 30,  and 37. Per above our fractional bandwidth is
        // (110-16)/110.
        let f = dir.frac_for_role(WeightRole::Exit, |u| u.rs.is_flagged_exit());
        assert!(((94.0 / 110.0) - f).abs() < 0.000001);
        // For middles: all relays are middles. We skipped 2, 9, 16,
        // 23, 30, and 37. Per above our fractional bandwidth is
        // (220-33)/220
        let f = dir.frac_for_role(WeightRole::Middle, |_| true);
        assert!(((187.0 / 220.0) - f).abs() < 0.000001);
        // Multiplying those together, we get the fraction of paths we can
        // build at ~0.64052066, which is above the threshold we set above for
        // MinPathsForCircsPct.
        let f = dir.frac_usable_paths();
        assert!((f - 0.64052066).abs() < 0.000001);
        // But if we try again with a slightly higher threshold...
        let mut dir = PartialNetDir::new(consensus, Some(&high_threshold));
        for (pos, md) in microdescs.into_iter().enumerate() {
            if pos % 7 == 2 {
                continue; // skip a few relays.
            }
            dir.add_microdesc(md);
        }
        assert!(dir.unwrap_if_sufficient().is_err());
    }
    /// Return a 3-tuple for use by `test_pick_*()` of an Rng, a number of
    /// iterations, and a tolerance.
    ///
    /// If the Rng is deterministic (the default), we can use a faster setup,
    /// with a higher tolerance and fewer iterations.  But if you've explicitly
    /// opted into randomization (or are replaying a seed from an earlier
    /// randomized test), we give you more iterations and a tighter tolerance.
    fn testing_rng_with_tolerances() -> (impl rand::Rng, usize, f64) {
        // Use a deterministic RNG if none is specified, since this is slow otherwise.
        let config = test_rng::Config::from_env().unwrap_or(test_rng::Config::Deterministic);
        let (iters, tolerance) = match config {
            test_rng::Config::Deterministic => (5000, 0.02),
            _ => (50000, 0.01),
        };
        (config.into_rng(), iters, tolerance)
    }
    #[test]
    fn test_pick() {
        let (consensus, microdescs) = construct_network().unwrap();
        let mut dir = PartialNetDir::new(consensus, None);
        for md in microdescs.into_iter() {
            let wanted = dir.add_microdesc(md.clone());
            assert!(wanted);
        }
        let dir = dir.unwrap_if_sufficient().unwrap();
        let (mut rng, total, tolerance) = testing_rng_with_tolerances();
        let mut picked = [0_isize; 40];
        for _ in 0..total {
            let r = dir.pick_relay(&mut rng, WeightRole::Middle, |r| {
                r.low_level_details().supports_exit_port_ipv4(80)
            });
            let r = r.unwrap();
            let id_byte = r.identity(RelayIdType::Rsa).unwrap().as_bytes()[0];
            picked[id_byte as usize] += 1;
        }
        // non-exits should never get picked.
        picked[0..10].iter().for_each(|x| assert_eq!(*x, 0));
        picked[20..30].iter().for_each(|x| assert_eq!(*x, 0));
        let picked_f: Vec<_> = picked.iter().map(|x| *x as f64 / total as f64).collect();
        // We didn't we any non-default weights, so the other relays get
        // weighted proportional to their bandwidth.
        assert_float_eq!(picked_f[19], (10.0 / 110.0), abs <= tolerance);
        assert_float_eq!(picked_f[38], (9.0 / 110.0), abs <= tolerance);
        assert_float_eq!(picked_f[39], (10.0 / 110.0), abs <= tolerance);
    }
    #[test]
    fn test_pick_multiple() {
        // This is mostly a copy of test_pick, except that it uses
        // pick_n_relays to pick several relays at once.
        let dir = construct_netdir().unwrap_if_sufficient().unwrap();
        let (mut rng, total, tolerance) = testing_rng_with_tolerances();
        let mut picked = [0_isize; 40];
        for _ in 0..total / 4 {
            let relays = dir.pick_n_relays(&mut rng, 4, WeightRole::Middle, |r| {
                r.low_level_details().supports_exit_port_ipv4(80)
            });
            assert_eq!(relays.len(), 4);
            for r in relays {
                let id_byte = r.identity(RelayIdType::Rsa).unwrap().as_bytes()[0];
                picked[id_byte as usize] += 1;
            }
        }
        // non-exits should never get picked.
        picked[0..10].iter().for_each(|x| assert_eq!(*x, 0));
        picked[20..30].iter().for_each(|x| assert_eq!(*x, 0));
        let picked_f: Vec<_> = picked.iter().map(|x| *x as f64 / total as f64).collect();
        // We didn't we any non-default weights, so the other relays get
        // weighted proportional to their bandwidth.
        assert_float_eq!(picked_f[19], (10.0 / 110.0), abs <= tolerance);
        assert_float_eq!(picked_f[36], (7.0 / 110.0), abs <= tolerance);
        assert_float_eq!(picked_f[39], (10.0 / 110.0), abs <= tolerance);
    }
    #[test]
    fn subnets() {
        let cfg = SubnetConfig::default();
        fn same_net(cfg: &SubnetConfig, a: &str, b: &str) -> bool {
            cfg.addrs_in_same_subnet(&a.parse().unwrap(), &b.parse().unwrap())
        }
        assert!(same_net(&cfg, "127.15.3.3", "127.15.9.9"));
        assert!(!same_net(&cfg, "127.15.3.3", "127.16.9.9"));
        assert!(!same_net(&cfg, "127.15.3.3", "127::"));
        assert!(same_net(&cfg, "ffff:ffff:90:33::", "ffff:ffff:91:34::"));
        assert!(!same_net(&cfg, "ffff:ffff:90:33::", "ffff:fffe:91:34::"));
        let cfg = SubnetConfig {
            subnets_family_v4: 32,
            subnets_family_v6: 128,
        };
        assert!(!same_net(&cfg, "127.15.3.3", "127.15.9.9"));
        assert!(!same_net(&cfg, "ffff:ffff:90:33::", "ffff:ffff:91:34::"));
        assert!(same_net(&cfg, "127.0.0.1", "127.0.0.1"));
        assert!(!same_net(&cfg, "127.0.0.1", "127.0.0.2"));
        assert!(same_net(&cfg, "ffff:ffff:90:33::", "ffff:ffff:90:33::"));
        let cfg = SubnetConfig {
            subnets_family_v4: 33,
            subnets_family_v6: 129,
        };
        assert!(!same_net(&cfg, "127.0.0.1", "127.0.0.1"));
        assert!(!same_net(&cfg, "::", "::"));
    }
    #[test]
    fn subnet_union() {
        let cfg1 = SubnetConfig {
            subnets_family_v4: 16,
            subnets_family_v6: 64,
        };
        let cfg2 = SubnetConfig {
            subnets_family_v4: 24,
            subnets_family_v6: 32,
        };
        let a1 = "1.2.3.4".parse().unwrap();
        let a2 = "1.2.10.10".parse().unwrap();
        let a3 = "ffff:ffff::7".parse().unwrap();
        let a4 = "ffff:ffff:1234::8".parse().unwrap();
        assert_eq!(cfg1.addrs_in_same_subnet(&a1, &a2), true);
        assert_eq!(cfg2.addrs_in_same_subnet(&a1, &a2), false);
        assert_eq!(cfg1.addrs_in_same_subnet(&a3, &a4), false);
        assert_eq!(cfg2.addrs_in_same_subnet(&a3, &a4), true);
        let cfg_u = cfg1.union(&cfg2);
        assert_eq!(
            cfg_u,
            SubnetConfig {
                subnets_family_v4: 16,
                subnets_family_v6: 32,
            }
        );
        assert_eq!(cfg_u.addrs_in_same_subnet(&a1, &a2), true);
        assert_eq!(cfg_u.addrs_in_same_subnet(&a3, &a4), true);
        assert_eq!(cfg1.union(&cfg1), cfg1);
        assert_eq!(cfg1.union(&SubnetConfig::no_addresses_match()), cfg1);
    }
    #[test]
    fn relay_funcs() {
        let (consensus, microdescs) = construct_custom_network(
            |pos, nb, _| {
                if pos == 15 {
                    nb.rs.add_or_port("[f0f0::30]:9001".parse().unwrap());
                } else if pos == 20 {
                    nb.rs.add_or_port("[f0f0::3131]:9001".parse().unwrap());
                }
            },
            None,
        )
        .unwrap();
        let subnet_config = SubnetConfig::default();
        let all_family_info = FamilyRules::all_family_info();
        let mut dir = PartialNetDir::new(consensus, None);
        for md in microdescs.into_iter() {
            let wanted = dir.add_microdesc(md.clone());
            assert!(wanted);
        }
        let dir = dir.unwrap_if_sufficient().unwrap();
        // Pick out a few relays by ID.
        let k0 = Ed25519Identity::from([0; 32]);
        let k1 = Ed25519Identity::from([1; 32]);
        let k2 = Ed25519Identity::from([2; 32]);
        let k3 = Ed25519Identity::from([3; 32]);
        let k10 = Ed25519Identity::from([10; 32]);
        let k15 = Ed25519Identity::from([15; 32]);
        let k20 = Ed25519Identity::from([20; 32]);
        let r0 = dir.by_id(&k0).unwrap();
        let r1 = dir.by_id(&k1).unwrap();
        let r2 = dir.by_id(&k2).unwrap();
        let r3 = dir.by_id(&k3).unwrap();
        let r10 = dir.by_id(&k10).unwrap();
        let r15 = dir.by_id(&k15).unwrap();
        let r20 = dir.by_id(&k20).unwrap();
        assert_eq!(r0.id(), &[0; 32].into());
        assert_eq!(r0.rsa_id(), &[0; 20].into());
        assert_eq!(r1.id(), &[1; 32].into());
        assert_eq!(r1.rsa_id(), &[1; 20].into());
        assert!(r0.same_relay_ids(&r0));
        assert!(r1.same_relay_ids(&r1));
        assert!(!r1.same_relay_ids(&r0));
        assert!(r0.low_level_details().is_dir_cache());
        assert!(!r1.low_level_details().is_dir_cache());
        assert!(r2.low_level_details().is_dir_cache());
        assert!(!r3.low_level_details().is_dir_cache());
        assert!(!r0.low_level_details().supports_exit_port_ipv4(80));
        assert!(!r1.low_level_details().supports_exit_port_ipv4(80));
        assert!(!r2.low_level_details().supports_exit_port_ipv4(80));
        assert!(!r3.low_level_details().supports_exit_port_ipv4(80));
        assert!(!r0.low_level_details().policies_allow_some_port());
        assert!(!r1.low_level_details().policies_allow_some_port());
        assert!(!r2.low_level_details().policies_allow_some_port());
        assert!(!r3.low_level_details().policies_allow_some_port());
        assert!(r10.low_level_details().policies_allow_some_port());
        assert!(r0.low_level_details().in_same_family(&r0, all_family_info));
        assert!(r0.low_level_details().in_same_family(&r1, all_family_info));
        assert!(r1.low_level_details().in_same_family(&r0, all_family_info));
        assert!(r1.low_level_details().in_same_family(&r1, all_family_info));
        assert!(!r0.low_level_details().in_same_family(&r2, all_family_info));
        assert!(!r2.low_level_details().in_same_family(&r0, all_family_info));
        assert!(r2.low_level_details().in_same_family(&r2, all_family_info));
        assert!(r2.low_level_details().in_same_family(&r3, all_family_info));
        assert!(r0.low_level_details().in_same_subnet(&r10, &subnet_config));
        assert!(r10.low_level_details().in_same_subnet(&r10, &subnet_config));
        assert!(r0.low_level_details().in_same_subnet(&r0, &subnet_config));
        assert!(r1.low_level_details().in_same_subnet(&r1, &subnet_config));
        assert!(!r1.low_level_details().in_same_subnet(&r2, &subnet_config));
        assert!(!r2.low_level_details().in_same_subnet(&r3, &subnet_config));
        // Make sure IPv6 families work.
        let subnet_config = SubnetConfig {
            subnets_family_v4: 128,
            subnets_family_v6: 96,
        };
        assert!(r15.low_level_details().in_same_subnet(&r20, &subnet_config));
        assert!(!r15.low_level_details().in_same_subnet(&r1, &subnet_config));
        // Make sure that subnet configs can be disabled.
        let subnet_config = SubnetConfig {
            subnets_family_v4: 255,
            subnets_family_v6: 255,
        };
        assert!(!r15.low_level_details().in_same_subnet(&r20, &subnet_config));
    }
    #[test]
    fn test_badexit() {
        // make a netdir where relays 10-19 are badexit, and everybody
        // exits to 443 on IPv6.
        use tor_netdoc::types::relay_flags::RelayFlag;
        let netdir = construct_custom_netdir(|pos, nb, _| {
            if (10..20).contains(&pos) {
                nb.rs.add_flags(RelayFlag::BadExit);
            }
            nb.md.parse_ipv6_policy("accept 443").unwrap();
        })
        .unwrap()
        .unwrap_if_sufficient()
        .unwrap();
        let e12 = netdir.by_id(&Ed25519Identity::from([12; 32])).unwrap();
        let e32 = netdir.by_id(&Ed25519Identity::from([32; 32])).unwrap();
        assert!(!e12.low_level_details().supports_exit_port_ipv4(80));
        assert!(e32.low_level_details().supports_exit_port_ipv4(80));
        assert!(!e12.low_level_details().supports_exit_port_ipv6(443));
        assert!(e32.low_level_details().supports_exit_port_ipv6(443));
        assert!(!e32.low_level_details().supports_exit_port_ipv6(555));
        assert!(!e12.low_level_details().policies_allow_some_port());
        assert!(e32.low_level_details().policies_allow_some_port());
        assert!(!e12.low_level_details().ipv4_policy().allows_some_port());
        assert!(!e12.low_level_details().ipv6_policy().allows_some_port());
        assert!(e32.low_level_details().ipv4_policy().allows_some_port());
        assert!(e32.low_level_details().ipv6_policy().allows_some_port());
        assert!(
            e12.low_level_details()
                .ipv4_declared_policy()
                .allows_some_port()
        );
        assert!(
            e12.low_level_details()
                .ipv6_declared_policy()
                .allows_some_port()
        );
    }
    #[cfg(feature = "experimental-api")]
    #[test]
    fn test_accessors() {
        let netdir = construct_netdir().unwrap_if_sufficient().unwrap();
        let r4 = netdir.by_id(&Ed25519Identity::from([4; 32])).unwrap();
        let r16 = netdir.by_id(&Ed25519Identity::from([16; 32])).unwrap();
        assert!(!r4.md().ipv4_policy().allows_some_port());
        assert!(r16.md().ipv4_policy().allows_some_port());
        assert!(!r4.rs().is_flagged_exit());
        assert!(r16.rs().is_flagged_exit());
    }
    #[test]
    fn test_by_id() {
        // Make a netdir that omits the microdescriptor for 0xDDDDDD...
        let netdir = construct_custom_netdir(|pos, nb, _| {
            nb.omit_md = pos == 13;
        })
        .unwrap();
        let netdir = netdir.unwrap_if_sufficient().unwrap();
        let r = netdir.by_id(&Ed25519Identity::from([0; 32])).unwrap();
        assert_eq!(r.id().as_bytes(), &[0; 32]);
        assert!(netdir.by_id(&Ed25519Identity::from([13; 32])).is_none());
        let r = netdir.by_rsa_id(&[12; 20].into()).unwrap();
        assert_eq!(r.rsa_id().as_bytes(), &[12; 20]);
        assert!(netdir.rsa_id_is_listed(&[12; 20].into()));
        assert!(netdir.by_rsa_id(&[13; 20].into()).is_none());
        assert!(netdir.by_rsa_id_unchecked(&[99; 20].into()).is_none());
        assert!(!netdir.rsa_id_is_listed(&[99; 20].into()));
        let r = netdir.by_rsa_id_unchecked(&[13; 20].into()).unwrap();
        assert_eq!(r.rs.rsa_identity().as_bytes(), &[13; 20]);
        assert!(netdir.rsa_id_is_listed(&[13; 20].into()));
        let pair_13_13 = RelayIds::builder()
            .ed_identity([13; 32].into())
            .rsa_identity([13; 20].into())
            .build()
            .unwrap();
        let pair_14_14 = RelayIds::builder()
            .ed_identity([14; 32].into())
            .rsa_identity([14; 20].into())
            .build()
            .unwrap();
        let pair_14_99 = RelayIds::builder()
            .ed_identity([14; 32].into())
            .rsa_identity([99; 20].into())
            .build()
            .unwrap();
        let r = netdir.by_ids(&pair_13_13);
        assert!(r.is_none());
        let r = netdir.by_ids(&pair_14_14).unwrap();
        assert_eq!(r.identity(RelayIdType::Rsa).unwrap().as_bytes(), &[14; 20]);
        assert_eq!(
            r.identity(RelayIdType::Ed25519).unwrap().as_bytes(),
            &[14; 32]
        );
        let r = netdir.by_ids(&pair_14_99);
        assert!(r.is_none());
        assert_eq!(
            netdir.id_pair_listed(&[13; 32].into(), &[13; 20].into()),
            None
        );
        assert_eq!(
            netdir.id_pair_listed(&[15; 32].into(), &[15; 20].into()),
            Some(true)
        );
        assert_eq!(
            netdir.id_pair_listed(&[15; 32].into(), &[99; 20].into()),
            Some(false)
        );
    }
    #[test]
    #[cfg(feature = "hs-common")]
    fn test_by_ids_detailed() {
        // Make a netdir that omits the microdescriptor for 0xDDDDDD...
        let netdir = construct_custom_netdir(|pos, nb, _| {
            nb.omit_md = pos == 13;
        })
        .unwrap();
        let netdir = netdir.unwrap_if_sufficient().unwrap();
        let id13_13 = RelayIds::builder()
            .ed_identity([13; 32].into())
            .rsa_identity([13; 20].into())
            .build()
            .unwrap();
        let id15_15 = RelayIds::builder()
            .ed_identity([15; 32].into())
            .rsa_identity([15; 20].into())
            .build()
            .unwrap();
        let id15_99 = RelayIds::builder()
            .ed_identity([15; 32].into())
            .rsa_identity([99; 20].into())
            .build()
            .unwrap();
        let id99_15 = RelayIds::builder()
            .ed_identity([99; 32].into())
            .rsa_identity([15; 20].into())
            .build()
            .unwrap();
        let id99_99 = RelayIds::builder()
            .ed_identity([99; 32].into())
            .rsa_identity([99; 20].into())
            .build()
            .unwrap();
        let id15_xx = RelayIds::builder()
            .ed_identity([15; 32].into())
            .build()
            .unwrap();
        let idxx_15 = RelayIds::builder()
            .rsa_identity([15; 20].into())
            .build()
            .unwrap();
        assert!(matches!(netdir.by_ids_detailed(&id13_13), Ok(None)));
        assert!(matches!(netdir.by_ids_detailed(&id15_15), Ok(Some(_))));
        assert!(matches!(
            netdir.by_ids_detailed(&id15_99),
            Err(RelayLookupError::Impossible)
        ));
        assert!(matches!(
            netdir.by_ids_detailed(&id99_15),
            Err(RelayLookupError::Impossible)
        ));
        assert!(matches!(netdir.by_ids_detailed(&id99_99), Ok(None)));
        assert!(matches!(netdir.by_ids_detailed(&id15_xx), Ok(Some(_))));
        assert!(matches!(netdir.by_ids_detailed(&idxx_15), Ok(Some(_))));
    }
    #[test]
    fn weight_type() {
        let r0 = RelayWeight(0);
        let r100 = RelayWeight(100);
        let r200 = RelayWeight(200);
        let r300 = RelayWeight(300);
        assert_eq!(r100 + r200, r300);
        assert_eq!(r100.checked_div(r200), Some(0.5));
        assert!(r100.checked_div(r0).is_none());
        assert_eq!(r200.ratio(0.5), Some(r100));
        assert!(r200.ratio(-1.0).is_none());
    }
    #[test]
    fn weight_accessors() {
        // Make a netdir that omits the microdescriptor for 0xDDDDDD...
        let netdir = construct_netdir().unwrap_if_sufficient().unwrap();
        let g_total = netdir.total_weight(WeightRole::Guard, |r| r.rs.is_flagged_guard());
        // This is just the total guard weight, since all our Wxy = 1.
        assert_eq!(g_total, RelayWeight(110_000));
        let g_total = netdir.total_weight(WeightRole::Guard, |_| false);
        assert_eq!(g_total, RelayWeight(0));
        let relay = netdir.by_id(&Ed25519Identity::from([35; 32])).unwrap();
        assert!(relay.rs.is_flagged_guard());
        let w = netdir.relay_weight(&relay, WeightRole::Guard);
        assert_eq!(w, RelayWeight(6_000));
        let w = netdir
            .weight_by_rsa_id(&[33; 20].into(), WeightRole::Guard)
            .unwrap();
        assert_eq!(w, RelayWeight(4_000));
        assert!(
            netdir
                .weight_by_rsa_id(&[99; 20].into(), WeightRole::Guard)
                .is_none()
        );
    }
    #[test]
    fn family_list() {
        let netdir = construct_custom_netdir(|pos, n, _| {
            if pos == 0x0a {
                n.md.family(
                    "$0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B0B \
                     $0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C0C \
                     $0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D0D"
                        .parse()
                        .unwrap(),
                );
            } else if pos == 0x0c {
                n.md.family("$0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A0A".parse().unwrap());
            }
        })
        .unwrap()
        .unwrap_if_sufficient()
        .unwrap();
        // In the testing netdir, adjacent members are in the same family by default...
        let r0 = netdir.by_id(&Ed25519Identity::from([0; 32])).unwrap();
        let family: Vec<_> = netdir.known_family_members(&r0).collect();
        assert_eq!(family.len(), 1);
        assert_eq!(family[0].id(), &Ed25519Identity::from([1; 32]));
        // But we've made this relay claim membership with several others.
        let r10 = netdir.by_id(&Ed25519Identity::from([10; 32])).unwrap();
        let family: HashSet<_> = netdir.known_family_members(&r10).map(|r| *r.id()).collect();
        assert_eq!(family.len(), 2);
        assert!(family.contains(&Ed25519Identity::from([11; 32])));
        assert!(family.contains(&Ed25519Identity::from([12; 32])));
        // Note that 13 doesn't get put in, even though it's listed, since it doesn't claim
        //  membership with 10.
    }
    #[test]
    #[cfg(feature = "geoip")]
    fn relay_has_country_code() {
        let src_v6 = r#"
        fe80:dead:beef::,fe80:dead:ffff::,US
        fe80:feed:eeee::1,fe80:feed:eeee::2,AT
        fe80:feed:eeee::2,fe80:feed:ffff::,DE
        "#;
        let db = GeoipDb::new_from_legacy_format("", src_v6).unwrap();
        let netdir = construct_custom_netdir_with_geoip(
            |pos, n, _| {
                if pos == 0x01 {
                    n.rs.add_or_port("[fe80:dead:beef::1]:42".parse().unwrap());
                }
                if pos == 0x02 {
                    n.rs.add_or_port("[fe80:feed:eeee::1]:42".parse().unwrap());
                    n.rs.add_or_port("[fe80:feed:eeee::2]:42".parse().unwrap());
                }
                if pos == 0x03 {
                    n.rs.add_or_port("[fe80:dead:beef::1]:42".parse().unwrap());
                    n.rs.add_or_port("[fe80:dead:beef::2]:42".parse().unwrap());
                }
            },
            &db,
        )
        .unwrap()
        .unwrap_if_sufficient()
        .unwrap();
        // No GeoIP data available -> None
        let r0 = netdir.by_id(&Ed25519Identity::from([0; 32])).unwrap();
        assert_eq!(r0.cc, None);
        // Exactly one match -> Some
        let r1 = netdir.by_id(&Ed25519Identity::from([1; 32])).unwrap();
        assert_eq!(r1.cc.as_ref().map(|x| x.as_ref()), Some("US"));
        // Conflicting matches -> None
        let r2 = netdir.by_id(&Ed25519Identity::from([2; 32])).unwrap();
        assert_eq!(r2.cc, None);
        // Multiple agreeing matches -> Some
        let r3 = netdir.by_id(&Ed25519Identity::from([3; 32])).unwrap();
        assert_eq!(r3.cc.as_ref().map(|x| x.as_ref()), Some("US"));
    }
    #[test]
    #[cfg(feature = "hs-common")]
    #[allow(deprecated)]
    fn hs_dirs_selection() {
        use tor_basic_utils::test_rng::testing_rng;
        const HSDIR_SPREAD_STORE: i32 = 6;
        const HSDIR_SPREAD_FETCH: i32 = 2;
        const PARAMS: [(&str, i32); 2] = [
            ("hsdir_spread_store", HSDIR_SPREAD_STORE),
            ("hsdir_spread_fetch", HSDIR_SPREAD_FETCH),
        ];
        let netdir: Arc<NetDir> =
            crate::testnet::construct_custom_netdir_with_params(|_, _, _| {}, PARAMS, None)
                .unwrap()
                .unwrap_if_sufficient()
                .unwrap()
                .into();
        let hsid = dummy_hs_blind_id();
        const OP_RELAY_COUNT: &[(HsDirOp, usize)] = &[
            // We can't upload to (hsdir_n_replicas * hsdir_spread_store) = 12, relays because there
            // are only 10 relays with the HsDir flag in the consensus.
            #[cfg(feature = "hs-service")]
            (HsDirOp::Upload, 10),
            (HsDirOp::Download, 4),
        ];
        for (op, relay_count) in OP_RELAY_COUNT {
            let relays = netdir.hs_dirs(&hsid, *op, &mut testing_rng());
            assert_eq!(relays.len(), *relay_count);
            // There should be no duplicates (the filtering function passed to
            // HsDirRing::ring_items_at() ensures the relays that are already in use for
            // lower-numbered replicas aren't considered a second time for a higher-numbered
            // replica).
            let unique = relays
                .iter()
                .map(|relay| relay.ed_identity())
                .collect::<HashSet<_>>();
            assert_eq!(unique.len(), relays.len());
        }
        // TODO: come up with a test that checks that HsDirRing::ring_items_at() skips over the
        // expected relays.
        //
        // For example, let's say we have the following hsdir ring:
        //
        //         A  -  B
        //        /       \
        //       F         C
        //        \       /
        //         E  -  D
        //
        // Let's also assume that:
        //
        //   * hsdir_spread_store = 3
        //   * the ordering of the relays on the ring is [A, B, C, D, E, F]
        //
        // If we use relays [A, B, C] for replica 1, and hs_index(2) = E, then replica 2 _must_ get
        // relays [E, F, D]. We should have a test that checks this.
    }
    #[test]
    fn zero_weights() {
        // Here we check the behavior of IndexedRandom::{choose_weighted, choose_multiple_weighted}
        // in the presence of items whose weight is 0.
        //
        // We think that the behavior is:
        //   - An item with weight 0 is never returned.
        //   - If all items have weight 0, choose_weighted returns an error.
        //   - If all items have weight 0, choose_multiple_weighted returns an empty list.
        //   - If we request n items from choose_multiple_weighted,
        //     but only m<n items have nonzero weight, we return all m of those items.
        //   - if the request for n items can't be completely satisfied with n items of weight >= 0,
        //     we get InsufficientNonZero.
        let items = vec![1, 2, 3];
        let mut rng = testing_rng();
        let a = items.choose_weighted(&mut rng, |_| 0);
        assert!(matches!(a, Err(WeightError::InsufficientNonZero)));
        let x = items.choose_multiple_weighted(&mut rng, 2, |_| 0);
        let xs: Vec<_> = x.unwrap().collect();
        assert!(xs.is_empty());
        let only_one = |n: &i32| if *n == 1 { 1 } else { 0 };
        let x = items.choose_multiple_weighted(&mut rng, 2, only_one);
        let xs: Vec<_> = x.unwrap().collect();
        assert_eq!(&xs[..], &[&1]);
        for _ in 0..100 {
            let a = items.choose_weighted(&mut rng, only_one);
            assert_eq!(a.unwrap(), &1);
            let x = items
                .choose_multiple_weighted(&mut rng, 1, only_one)
                .unwrap()
                .collect::<Vec<_>>();
            assert_eq!(x, vec![&1]);
        }
    }
    #[test]
    fn insufficient_but_nonzero() {
        // Here we check IndexedRandom::choose_multiple_weighted when there no zero values,
        // but there are insufficient values.
        // (If this behavior changes, we need to change our usage.)
        let items = vec![1, 2, 3];
        let mut rng = testing_rng();
        let mut a = items
            .choose_multiple_weighted(&mut rng, 10, |_| 1)
            .unwrap()
            .copied()
            .collect::<Vec<_>>();
        a.sort();
        assert_eq!(a, items);
    }
}