clippy: drop unneeded returns

This commit is contained in:
Graham Christensen 2019-01-02 19:38:03 -05:00
parent 9b9941f7e6
commit 1b5287d6b0
No known key found for this signature in database
GPG key ID: ACA1C1D120C83D5C
31 changed files with 175 additions and 176 deletions

View file

@ -615,7 +615,7 @@ impl MetricCollector {
f.write_all(variants.join("\n").as_bytes()).unwrap(); f.write_all(variants.join("\n").as_bytes()).unwrap();
f.write_all("return output;\n }".as_bytes()).unwrap(); f.write_all("output\n }".as_bytes()).unwrap();
f.write_all("\n}".as_bytes()).unwrap(); f.write_all("\n}".as_bytes()).unwrap();
} }

View file

@ -13,11 +13,12 @@ impl ACL {
) -> ACL { ) -> ACL {
trusted_users.iter_mut().map(|x| *x = x.to_lowercase()).last(); trusted_users.iter_mut().map(|x| *x = x.to_lowercase()).last();
known_users.iter_mut().map(|x| *x = x.to_lowercase()).last(); known_users.iter_mut().map(|x| *x = x.to_lowercase()).last();
return ACL {
ACL {
trusted_users, trusted_users,
known_users, known_users,
repos, repos,
}; }
} }
pub fn is_repo_eligible(&self, name: &str) -> bool { pub fn is_repo_eligible(&self, name: &str) -> bool {
@ -46,16 +47,16 @@ impl ACL {
return false; return false;
} }
return self.known_users.contains(&user.to_lowercase()); self.known_users.contains(&user.to_lowercase())
} }
pub fn can_build_unrestricted(&self, user: &str, repo: &str) -> bool { pub fn can_build_unrestricted(&self, user: &str, repo: &str) -> bool {
if repo.to_lowercase() == "nixos/nixpkgs" { if repo.to_lowercase() == "nixos/nixpkgs" {
return self.trusted_users.contains(&user.to_lowercase()); self.trusted_users.contains(&user.to_lowercase())
} else if user == "grahamc" { } else if user == "grahamc" {
return true; true
} else { } else {
return false; false
} }
} }
} }

View file

@ -177,7 +177,7 @@ impl AsyncCmd {
waiters.len() waiters.len()
); );
return return_status; return_status
}); });
SpawnedAsyncCmd { SpawnedAsyncCmd {

View file

@ -32,5 +32,5 @@ fn main() {
fn file_to_str(f: &mut File) -> String { fn file_to_str(f: &mut File) -> String {
let mut buffer = Vec::new(); let mut buffer = Vec::new();
f.read_to_end(&mut buffer).expect("Reading eval output"); f.read_to_end(&mut buffer).expect("Reading eval output");
return String::from(String::from_utf8_lossy(&buffer)); String::from(String::from_utf8_lossy(&buffer))
} }

View file

@ -13,7 +13,7 @@ pub struct CachedCloner {
} }
pub fn cached_cloner(path: &Path) -> CachedCloner { pub fn cached_cloner(path: &Path) -> CachedCloner {
return CachedCloner { root: path.to_path_buf() }; CachedCloner { root: path.to_path_buf() }
} }
pub struct CachedProject { pub struct CachedProject {
@ -39,10 +39,10 @@ impl CachedCloner {
new_root.push("repo"); new_root.push("repo");
new_root.push(format!("{:x}", md5::compute(&name))); new_root.push(format!("{:x}", md5::compute(&name)));
return CachedProject { CachedProject {
root: new_root, root: new_root,
clone_url, clone_url,
}; }
} }
} }
@ -53,12 +53,12 @@ impl CachedProject {
let mut new_root = self.root.clone(); let mut new_root = self.root.clone();
new_root.push(use_category); new_root.push(use_category);
return Ok(CachedProjectCo { Ok(CachedProjectCo {
root: new_root, root: new_root,
id, id,
clone_url: self.clone_from().clone(), clone_url: self.clone_from().clone(),
local_reference: self.clone_to().clone(), local_reference: self.clone_to().clone(),
}); })
} }
fn prefetch_cache(&self) -> Result<PathBuf, Error> { fn prefetch_cache(&self) -> Result<PathBuf, Error> {
@ -67,7 +67,7 @@ impl CachedProject {
self.clone_repo()?; self.clone_repo()?;
self.fetch_repo()?; self.fetch_repo()?;
return Ok(self.clone_to()); Ok(self.clone_to())
} }
} }
@ -89,7 +89,7 @@ impl CachedProjectCo {
// let build_dir = self.build_dir(); // let build_dir = self.build_dir();
return Ok(self.clone_to().to_str().unwrap().to_string()); Ok(self.clone_to().to_str().unwrap().to_string())
} }
pub fn fetch_pr(&self, pr_id: u64) -> Result<(), Error> { pub fn fetch_pr(&self, pr_id: u64) -> Result<(), Error> {
@ -105,9 +105,9 @@ impl CachedProjectCo {
lock.unlock(); lock.unlock();
if result.success() { if result.success() {
return Ok(()); Ok(())
} else { } else {
return Err(Error::new(ErrorKind::Other, "Failed to fetch PR")); Err(Error::new(ErrorKind::Other, "Failed to fetch PR"))
} }
} }
@ -124,7 +124,7 @@ impl CachedProjectCo {
lock.unlock(); lock.unlock();
return result.success(); result.success()
} }
pub fn merge_commit(&self, commit: &OsStr) -> Result<(), Error> { pub fn merge_commit(&self, commit: &OsStr) -> Result<(), Error> {
@ -142,9 +142,9 @@ impl CachedProjectCo {
lock.unlock(); lock.unlock();
if result.success() { if result.success() {
return Ok(()); Ok(())
} else { } else {
return Err(Error::new(ErrorKind::Other, "Failed to merge")); Err(Error::new(ErrorKind::Other, "Failed to merge"))
} }
} }
@ -161,17 +161,17 @@ impl CachedProjectCo {
lock.unlock(); lock.unlock();
if result.status.success() { if result.status.success() {
return Ok( Ok(
String::from_utf8_lossy(&result.stdout) String::from_utf8_lossy(&result.stdout)
.lines() .lines()
.map(|l| l.to_owned()) .map(|l| l.to_owned())
.collect(), .collect(),
); )
} else { } else {
return Err(Error::new( Err(Error::new(
ErrorKind::Other, ErrorKind::Other,
String::from_utf8_lossy(&result.stderr).to_lowercase(), String::from_utf8_lossy(&result.stderr).to_lowercase(),
)); ))
} }
} }
@ -188,67 +188,67 @@ impl CachedProjectCo {
lock.unlock(); lock.unlock();
if result.status.success() { if result.status.success() {
return Ok( Ok(
String::from_utf8_lossy(&result.stdout) String::from_utf8_lossy(&result.stdout)
.lines() .lines()
.map(|l| l.to_owned()) .map(|l| l.to_owned())
.collect(), .collect(),
); )
} else { } else {
return Err(Error::new( Err(Error::new(
ErrorKind::Other, ErrorKind::Other,
String::from_utf8_lossy(&result.stderr).to_lowercase(), String::from_utf8_lossy(&result.stderr).to_lowercase(),
)); ))
} }
} }
} }
impl clone::GitClonable for CachedProjectCo { impl clone::GitClonable for CachedProjectCo {
fn clone_from(&self) -> String { fn clone_from(&self) -> String {
return self.clone_url.clone(); self.clone_url.clone()
} }
fn clone_to(&self) -> PathBuf { fn clone_to(&self) -> PathBuf {
let mut clone_path = self.root.clone(); let mut clone_path = self.root.clone();
clone_path.push(&self.id); clone_path.push(&self.id);
return clone_path; clone_path
} }
fn lock_path(&self) -> PathBuf { fn lock_path(&self) -> PathBuf {
let mut lock_path = self.root.clone(); let mut lock_path = self.root.clone();
lock_path.push(format!("{}.lock", self.id)); lock_path.push(format!("{}.lock", self.id));
return lock_path; lock_path
} }
fn extra_clone_args(&self) -> Vec<&OsStr> { fn extra_clone_args(&self) -> Vec<&OsStr> {
let local_ref = self.local_reference.as_ref(); let local_ref = self.local_reference.as_ref();
return vec![ vec![
OsStr::new("--shared"), OsStr::new("--shared"),
OsStr::new("--reference-if-able"), OsStr::new("--reference-if-able"),
local_ref, local_ref,
]; ]
} }
} }
impl clone::GitClonable for CachedProject { impl clone::GitClonable for CachedProject {
fn clone_from(&self) -> String { fn clone_from(&self) -> String {
return self.clone_url.clone(); self.clone_url.clone()
} }
fn clone_to(&self) -> PathBuf { fn clone_to(&self) -> PathBuf {
let mut clone_path = self.root.clone(); let mut clone_path = self.root.clone();
clone_path.push("clone"); clone_path.push("clone");
return clone_path; clone_path
} }
fn lock_path(&self) -> PathBuf { fn lock_path(&self) -> PathBuf {
let mut clone_path = self.root.clone(); let mut clone_path = self.root.clone();
clone_path.push("clone.lock"); clone_path.push("clone.lock");
return clone_path; clone_path
} }
fn extra_clone_args(&self) -> Vec<&OsStr> { fn extra_clone_args(&self) -> Vec<&OsStr> {
return vec![OsStr::new("--bare")]; vec![OsStr::new("--bare")]
} }
} }

View file

@ -28,7 +28,7 @@ pub trait GitClonable {
match fs::File::create(self.lock_path()) { match fs::File::create(self.lock_path()) {
Err(e) => { Err(e) => {
warn!("Failed to create lock file {:?}: {}", self.lock_path(), e); warn!("Failed to create lock file {:?}: {}", self.lock_path(), e);
return Err(e); Err(e)
} }
Ok(lock) => { Ok(lock) => {
match lock.lock_exclusive() { match lock.lock_exclusive() {
@ -38,11 +38,11 @@ pub trait GitClonable {
self.lock_path(), self.lock_path(),
e e
); );
return Err(e); Err(e)
} }
Ok(_) => { Ok(_) => {
debug!("Got lock on {:?}", self.lock_path()); debug!("Got lock on {:?}", self.lock_path());
return Ok(Lock { lock: Some(lock) }); Ok(Lock { lock: Some(lock) })
} }
} }
} }
@ -74,9 +74,9 @@ pub trait GitClonable {
lock.unlock(); lock.unlock();
if result.success() { if result.success() {
return Ok(()); Ok(())
} else { } else {
return Err(Error::new(ErrorKind::Other, format!("Failed to clone from {:?} to {:?}", self.clone_from(), self.clone_to()))); Err(Error::new(ErrorKind::Other, format!("Failed to clone from {:?} to {:?}", self.clone_from(), self.clone_to())))
} }
} }
@ -93,9 +93,9 @@ pub trait GitClonable {
lock.unlock(); lock.unlock();
if result.success() { if result.success() {
return Ok(()); Ok(())
} else { } else {
return Err(Error::new(ErrorKind::Other, "Failed to fetch")); Err(Error::new(ErrorKind::Other, "Failed to fetch"))
} }
} }
@ -125,7 +125,7 @@ pub trait GitClonable {
lock.unlock(); lock.unlock();
return Ok(()); Ok(())
} }
fn checkout(&self, git_ref: &OsStr) -> Result<(), Error> { fn checkout(&self, git_ref: &OsStr) -> Result<(), Error> {
@ -142,9 +142,9 @@ pub trait GitClonable {
lock.unlock(); lock.unlock();
if result.success() { if result.success() {
return Ok(()); Ok(())
} else { } else {
return Err(Error::new(ErrorKind::Other, "Failed to checkout")); Err(Error::new(ErrorKind::Other, "Failed to checkout"))
} }
} }
} }

View file

@ -9,9 +9,9 @@ pub fn parse(text: &str) -> Option<Vec<Instruction>> {
.collect(); .collect();
if instructions.len() == 0 { if instructions.len() == 0 {
return None; None
} else { } else {
return Some(instructions); Some(instructions)
} }
} }

View file

@ -29,7 +29,7 @@ impl<'a> CommitStatus<'a> {
stat.set_url(url); stat.set_url(url);
return stat; stat
} }
pub fn set_url(&mut self, url: Option<String>) { pub fn set_url(&mut self, url: Option<String>) {

View file

@ -87,11 +87,11 @@ pub struct CheckoutConfig {
impl Config { impl Config {
pub fn whoami(&self) -> String { pub fn whoami(&self) -> String {
return format!("{}-{}", self.runner.identity, self.nix.system); format!("{}-{}", self.runner.identity, self.nix.system)
} }
pub fn acl(&self) -> acl::ACL { pub fn acl(&self) -> acl::ACL {
return acl::ACL::new( acl::ACL::new(
self.runner.repos.clone().expect( self.runner.repos.clone().expect(
"fetching config's runner.repos", "fetching config's runner.repos",
), ),
@ -101,7 +101,7 @@ impl Config {
self.runner.known_users.clone().expect( self.runner.known_users.clone().expect(
"fetching config's runner.known_users", "fetching config's runner.known_users",
), ),
); )
} }
pub fn github(&self) -> Github { pub fn github(&self) -> Github {
@ -141,25 +141,25 @@ impl Config {
panic!(); panic!();
} }
return Nix::new( Nix::new(
self.nix.system.clone(), self.nix.system.clone(),
self.nix.remote.clone(), self.nix.remote.clone(),
self.nix.build_timeout_seconds, self.nix.build_timeout_seconds,
self.nix.initial_heap_size.clone(), self.nix.initial_heap_size.clone(),
); )
} }
} }
impl RabbitMQConfig { impl RabbitMQConfig {
pub fn as_uri(&self) -> String { pub fn as_uri(&self) -> String {
return format!( format!(
"{}://{}:{}@{}/{}", "{}://{}:{}@{}/{}",
if self.ssl { "amqps" } else { "amqp" }, if self.ssl { "amqps" } else { "amqp" },
self.username, self.username,
self.password, self.password,
self.host, self.host,
self.virtualhost.clone().unwrap_or("/".to_owned()), self.virtualhost.clone().unwrap_or("/".to_owned()),
); )
} }
} }
@ -170,5 +170,5 @@ pub fn load(filename: &Path) -> Config {
let deserialized: Config = serde_json::from_str(&contents).unwrap(); let deserialized: Config = serde_json::from_str(&contents).unwrap();
return deserialized; deserialized
} }

View file

@ -312,7 +312,7 @@ pub fn session_from_config(config: &RabbitMQConfig) -> Result<amqp::Session, amq
let session = try!(amqp::Session::new(options)); let session = try!(amqp::Session::new(options));
info!("Connected to {}", &config.host); info!("Connected to {}", &config.host);
return Ok(session); Ok(session)
} }
pub trait TypedWrappers { pub trait TypedWrappers {

View file

@ -33,6 +33,6 @@ impl EvalChecker {
pub fn cli_cmd(&self) -> String { pub fn cli_cmd(&self) -> String {
let mut cli = vec![self.op.to_string()]; let mut cli = vec![self.op.to_string()];
cli.append(&mut self.args.clone()); cli.append(&mut self.args.clone());
return cli.join(" "); cli.join(" ")
} }
} }

View file

@ -4,5 +4,5 @@ use std::fs::File;
pub fn file_to_str(f: &mut File) -> String { pub fn file_to_str(f: &mut File) -> String {
let mut buffer = Vec::new(); let mut buffer = Vec::new();
f.read_to_end(&mut buffer).expect("Reading eval output"); f.read_to_end(&mut buffer).expect("Reading eval output");
return String::from(String::from_utf8_lossy(&buffer)); String::from(String::from_utf8_lossy(&buffer))
} }

View file

@ -75,7 +75,7 @@ pub mod ofborg {
pub use test_scratch; pub use test_scratch;
pub use easyamqp; pub use easyamqp;
pub const VERSION: &'static str = env!("CARGO_PKG_VERSION"); pub const VERSION: &str = env!("CARGO_PKG_VERSION");
pub fn partition_result<A,B>(results: Vec<Result<A,B>>) -> (Vec<A>, Vec<B>) { pub fn partition_result<A,B>(results: Vec<Result<A,B>>) -> (Vec<A>, Vec<B>) {
let mut ok = Vec::new(); let mut ok = Vec::new();

View file

@ -10,7 +10,7 @@ pub trait Lockable {
fn lock(&self) -> Result<Lock, Error> { fn lock(&self) -> Result<Lock, Error> {
let lock = fs::File::create(self.lock_path())?; let lock = fs::File::create(self.lock_path())?;
lock.lock_exclusive()?; lock.lock_exclusive()?;
return Ok(Lock { lock: Some(lock) }); Ok(Lock { lock: Some(lock) })
} }
} }

View file

@ -49,7 +49,7 @@ impl BuildJob {
} }
pub fn from(data: &Vec<u8>) -> Result<BuildJob, serde_json::error::Error> { pub fn from(data: &Vec<u8>) -> Result<BuildJob, serde_json::error::Error> {
return serde_json::from_slice(&data); serde_json::from_slice(&data)
} }
pub struct Actions { pub struct Actions {

View file

@ -4,7 +4,7 @@ use serde_json;
pub fn from(data: &Vec<u8>) -> Result<MassRebuildJob, serde_json::error::Error> { pub fn from(data: &Vec<u8>) -> Result<MassRebuildJob, serde_json::error::Error> {
return serde_json::from_slice(&data); serde_json::from_slice(&data)
} }
#[derive(Serialize, Deserialize, Debug)] #[derive(Serialize, Deserialize, Debug)]
@ -17,7 +17,7 @@ pub struct Actions {}
impl Actions { impl Actions {
pub fn skip(&mut self, _job: &MassRebuildJob) -> worker::Actions { pub fn skip(&mut self, _job: &MassRebuildJob) -> worker::Actions {
return vec![worker::Action::Ack]; vec![worker::Action::Ack]
} }
pub fn done( pub fn done(
@ -26,6 +26,6 @@ impl Actions {
mut response: worker::Actions, mut response: worker::Actions,
) -> worker::Actions { ) -> worker::Actions {
response.push(worker::Action::Ack); response.push(worker::Action::Ack);
return response; response
} }
} }

View file

@ -75,31 +75,31 @@ pub struct Nix {
impl Nix { impl Nix {
pub fn new(system: String, remote: String, build_timeout: u16, initial_heap_size: Option<String>) -> Nix { pub fn new(system: String, remote: String, build_timeout: u16, initial_heap_size: Option<String>) -> Nix {
return Nix { Nix {
system, system,
remote, remote,
build_timeout, build_timeout,
initial_heap_size, initial_heap_size,
limit_supported_systems: true, limit_supported_systems: true,
}; }
} }
pub fn with_system(&self, system: String) -> Nix { pub fn with_system(&self, system: String) -> Nix {
let mut n = self.clone(); let mut n = self.clone();
n.system = system; n.system = system;
return n; n
} }
pub fn with_limited_supported_systems(&self) -> Nix { pub fn with_limited_supported_systems(&self) -> Nix {
let mut n = self.clone(); let mut n = self.clone();
n.limit_supported_systems = true; n.limit_supported_systems = true;
return n; n
} }
pub fn without_limited_supported_systems(&self) -> Nix { pub fn without_limited_supported_systems(&self) -> Nix {
let mut n = self.clone(); let mut n = self.clone();
n.limit_supported_systems = false; n.limit_supported_systems = false;
return n; n
} }
pub fn safely_partition_instantiable_attrs( pub fn safely_partition_instantiable_attrs(
@ -134,7 +134,7 @@ impl Nix {
) -> Result<File, File> { ) -> Result<File, File> {
let cmd = self.safely_instantiate_attrs_cmd(nixpkgs, file, attrs); let cmd = self.safely_instantiate_attrs_cmd(nixpkgs, file, attrs);
return self.run(cmd, true); self.run(cmd, true)
} }
pub fn safely_instantiate_attrs_cmd( pub fn safely_instantiate_attrs_cmd(
@ -150,7 +150,7 @@ impl Nix {
attrargs.push(attr); attrargs.push(attr);
} }
return self.safe_command(Operation::Instantiate, nixpkgs, attrargs); self.safe_command(Operation::Instantiate, nixpkgs, attrargs)
} }
pub fn safely_build_attrs( pub fn safely_build_attrs(
@ -161,7 +161,7 @@ impl Nix {
) -> Result<File, File> { ) -> Result<File, File> {
let cmd = self.safely_build_attrs_cmd(nixpkgs, file, attrs); let cmd = self.safely_build_attrs_cmd(nixpkgs, file, attrs);
return self.run(cmd, true); self.run(cmd, true)
} }
pub fn safely_build_attrs_async( pub fn safely_build_attrs_async(
@ -197,7 +197,7 @@ impl Nix {
args: Vec<String>, args: Vec<String>,
keep_stdout: bool, keep_stdout: bool,
) -> Result<File, File> { ) -> Result<File, File> {
return self.run(self.safe_command(op, nixpkgs, args), keep_stdout); self.run(self.safe_command(op, nixpkgs, args), keep_stdout)
} }
pub fn run(&self, mut cmd: Command, keep_stdout: bool) -> Result<File, File> { pub fn run(&self, mut cmd: Command, keep_stdout: bool) -> Result<File, File> {
@ -223,9 +223,9 @@ impl Nix {
); );
if status.success() { if status.success() {
return Ok(reader); Ok(reader)
} else { } else {
return Err(reader); Err(reader)
} }
} }
@ -271,7 +271,7 @@ impl Nix {
command.args(args); command.args(args);
return command; command
} }
} }
@ -302,19 +302,19 @@ mod tests {
fn build_path() -> PathBuf { fn build_path() -> PathBuf {
let mut cwd = env::current_dir().unwrap(); let mut cwd = env::current_dir().unwrap();
cwd.push(Path::new("./test-srcs/build")); cwd.push(Path::new("./test-srcs/build"));
return cwd; cwd
} }
fn passing_eval_path() -> PathBuf { fn passing_eval_path() -> PathBuf {
let mut cwd = env::current_dir().unwrap(); let mut cwd = env::current_dir().unwrap();
cwd.push(Path::new("./test-srcs/eval")); cwd.push(Path::new("./test-srcs/eval"));
return cwd; cwd
} }
fn individual_eval_path() -> PathBuf { fn individual_eval_path() -> PathBuf {
let mut cwd = env::current_dir().unwrap(); let mut cwd = env::current_dir().unwrap();
cwd.push(Path::new("./test-srcs/eval-mixed-failure")); cwd.push(Path::new("./test-srcs/eval-mixed-failure"));
return cwd; cwd
} }
fn strip_ansi(string: &str) -> String { fn strip_ansi(string: &str) -> String {

View file

@ -48,10 +48,10 @@ pub struct ChannelNotificationReceiver<'a> {
impl<'a> ChannelNotificationReceiver<'a> { impl<'a> ChannelNotificationReceiver<'a> {
pub fn new(channel: &'a mut Channel, delivery_tag: u64) -> ChannelNotificationReceiver<'a> { pub fn new(channel: &'a mut Channel, delivery_tag: u64) -> ChannelNotificationReceiver<'a> {
return ChannelNotificationReceiver { ChannelNotificationReceiver {
channel, channel,
delivery_tag, delivery_tag,
}; }
} }
} }
@ -87,7 +87,7 @@ impl<'a> NotificationReceiver for ChannelNotificationReceiver<'a> {
} }
pub fn new<T: SimpleNotifyWorker>(worker: T) -> NotifyWorker<T> { pub fn new<T: SimpleNotifyWorker>(worker: T) -> NotifyWorker<T> {
return NotifyWorker { internal: worker }; NotifyWorker { internal: worker }
} }
impl<T: SimpleNotifyWorker + Send> Consumer for NotifyWorker<T> { impl<T: SimpleNotifyWorker + Send> Consumer for NotifyWorker<T> {

View file

@ -30,11 +30,11 @@ impl OutPathDiff {
match x { match x {
Ok(f) => { Ok(f) => {
self.original = Some(f); self.original = Some(f);
return Ok(true); Ok(true)
} }
Err(e) => { Err(e) => {
info!("Failed to find Before list"); info!("Failed to find Before list");
return Err(e); Err(e)
} }
} }
} }
@ -49,11 +49,11 @@ impl OutPathDiff {
match x { match x {
Ok(f) => { Ok(f) => {
self.current = Some(f); self.current = Some(f);
return Ok(true); Ok(true)
} }
Err(e) => { Err(e) => {
info!("Failed to find After list"); info!("Failed to find After list");
return Err(e); Err(e)
} }
} }
} }
@ -66,12 +66,12 @@ impl OutPathDiff {
let removed: Vec<PackageArch> = orig_set.difference(&cur_set).map(|ref p| (**p).clone()).collect(); let removed: Vec<PackageArch> = orig_set.difference(&cur_set).map(|ref p| (**p).clone()).collect();
let added: Vec<PackageArch> = cur_set.difference(&orig_set).map(|ref p| (**p).clone()).collect(); let added: Vec<PackageArch> = cur_set.difference(&orig_set).map(|ref p| (**p).clone()).collect();
return Some((removed, added)); Some((removed, added))
} else { } else {
return None; None
} }
} else { } else {
return None; None
} }
} }
@ -94,7 +94,7 @@ impl OutPathDiff {
} }
} }
return None; None
} }
fn run(&mut self) -> Result<PackageOutPaths, File> { fn run(&mut self) -> Result<PackageOutPaths, File> {

View file

@ -18,7 +18,7 @@ impl StdenvTagger {
}; };
t.possible.sort(); t.possible.sort();
return t; t
} }
pub fn changed(&mut self, systems: Vec<tasks::eval::stdenvs::System>) { pub fn changed(&mut self, systems: Vec<tasks::eval::stdenvs::System>) {
@ -54,7 +54,7 @@ impl StdenvTagger {
remove.remove(pos); remove.remove(pos);
} }
return remove; remove
} }
} }
@ -74,7 +74,7 @@ impl PkgsAddedRemovedTagger {
}; };
t.possible.sort(); t.possible.sort();
return t; t
} }
pub fn changed(&mut self, removed: Vec<PackageArch>, added: Vec<PackageArch>) { pub fn changed(&mut self, removed: Vec<PackageArch>, added: Vec<PackageArch>) {
@ -93,7 +93,7 @@ impl PkgsAddedRemovedTagger {
pub fn tags_to_remove(&self) -> Vec<String> { pub fn tags_to_remove(&self) -> Vec<String> {
// The cleanup tag is too vague to automatically remove. // The cleanup tag is too vague to automatically remove.
return vec![]; vec![]
} }
} }
@ -122,7 +122,7 @@ impl RebuildTagger {
}; };
t.possible.sort(); t.possible.sort();
return t; t
} }
pub fn parse_attrs(&mut self, attrs: Vec<PackageArch>) { pub fn parse_attrs(&mut self, attrs: Vec<PackageArch>) {
@ -175,22 +175,21 @@ impl RebuildTagger {
remove.remove(pos); remove.remove(pos);
} }
return remove; remove
} }
fn bucket(&self, count: u64) -> &str { fn bucket(&self, count: u64) -> &str {
if count > 500 { if count > 500 {
return "501+"; "501+"
} else if count > 100 { } else if count > 100 {
return "101-500"; "101-500"
} else if count > 10 { } else if count > 10 {
return "11-100"; "11-100"
} else if count > 0 { } else if count > 0 {
return "1-10"; "1-10"
} else { } else {
return "0"; "0"
} }
} }
} }
@ -232,7 +231,7 @@ impl PathsTagger {
remove.remove(pos); remove.remove(pos);
} }
return remove; remove
} }
} }

View file

@ -31,12 +31,12 @@ impl BuildWorker {
system: String, system: String,
identity: String, identity: String,
) -> BuildWorker { ) -> BuildWorker {
return BuildWorker { BuildWorker {
cloner, cloner,
nix, nix,
system, system,
identity, identity,
}; }
} }
fn actions<'a, 'b>( fn actions<'a, 'b>(
@ -80,7 +80,7 @@ impl<'a, 'b> JobActions<'a, 'b> {
None, None,
)); ));
return JobActions { JobActions {
system: system.to_owned(), system: system.to_owned(),
identity: identity.to_owned(), identity: identity.to_owned(),
receiver, receiver,
@ -92,7 +92,7 @@ impl<'a, 'b> JobActions<'a, 'b> {
log_routing_key, log_routing_key,
result_exchange, result_exchange,
result_routing_key, result_routing_key,
}; }
} }
pub fn log_snippet(&self) -> Vec<String> { pub fn log_snippet(&self) -> Vec<String> {
@ -128,7 +128,6 @@ impl<'a, 'b> JobActions<'a, 'b> {
let result_exchange = self.result_exchange.clone(); let result_exchange = self.result_exchange.clone();
let result_routing_key = self.result_routing_key.clone(); let result_routing_key = self.result_routing_key.clone();
self.tell(worker::publish_serde_action( self.tell(worker::publish_serde_action(
result_exchange, result_exchange,
result_routing_key, result_routing_key,
@ -280,13 +279,13 @@ impl notifyworker::SimpleNotifyWorker for BuildWorker {
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
println!("lmao I got a job?"); println!("lmao I got a job?");
return match buildjob::from(body) { match buildjob::from(body) {
Ok(e) => Ok(e), Ok(e) => Ok(e),
Err(e) => { Err(e) => {
println!("{:?}", String::from_utf8(body.clone())); println!("{:?}", String::from_utf8(body.clone()));
panic!("{:?}", e); panic!("{:?}", e);
} }
}; }
} }
fn consumer( fn consumer(
@ -421,7 +420,7 @@ mod tests {
} }
fn tpath(component: &str) -> PathBuf { fn tpath(component: &str) -> PathBuf {
return Path::new(env!("CARGO_MANIFEST_DIR")).join(component); Path::new(env!("CARGO_MANIFEST_DIR")).join(component)
} }
fn make_worker(path: &Path) -> BuildWorker { fn make_worker(path: &Path) -> BuildWorker {
@ -434,7 +433,7 @@ mod tests {
"cargo-test-build".to_owned(), "cargo-test-build".to_owned(),
); );
return worker; worker
} }
fn make_pr_repo(bare: &Path, co: &Path) -> String { fn make_pr_repo(bare: &Path, co: &Path) -> String {
@ -447,7 +446,8 @@ mod tests {
.output() .output()
.expect("building the test PR failed"); .expect("building the test PR failed");
let hash = String::from_utf8(output.stdout).expect("Should just be a hash"); let hash = String::from_utf8(output.stdout).expect("Should just be a hash");
return hash.trim().to_owned();
hash.trim().to_owned()
} }
fn strip_escaped_ansi(string: &str) -> String { fn strip_escaped_ansi(string: &str) -> String {

View file

@ -28,7 +28,7 @@ pub struct Stdenvs {
impl Stdenvs { impl Stdenvs {
pub fn new(nix: nix::Nix, co: PathBuf) -> Stdenvs { pub fn new(nix: nix::Nix, co: PathBuf) -> Stdenvs {
return Stdenvs { Stdenvs {
nix, nix,
co, co,
@ -37,7 +37,7 @@ impl Stdenvs {
darwin_stdenv_before: None, darwin_stdenv_before: None,
darwin_stdenv_after: None, darwin_stdenv_after: None,
}; }
} }
pub fn identify_before(&mut self) { pub fn identify_before(&mut self) {
@ -51,7 +51,7 @@ impl Stdenvs {
} }
pub fn are_same(&self) -> bool { pub fn are_same(&self) -> bool {
return self.changed().len() == 0; self.changed().len() == 0
} }
pub fn changed(&self) -> Vec<System> { pub fn changed(&self) -> Vec<System> {
@ -65,8 +65,7 @@ impl Stdenvs {
changed.push(System::X8664Darwin); changed.push(System::X8664Darwin);
} }
changed
return changed;
} }
fn identify(&mut self, system: System, from: StdenvFrom) { fn identify(&mut self, system: System, from: StdenvFrom) {
@ -104,13 +103,13 @@ impl Stdenvs {
println!("{:?}", result); println!("{:?}", result);
return match result { match result {
Ok(mut out) => Some(file_to_str(&mut out)), Ok(mut out) => Some(file_to_str(&mut out)),
Err(mut out) => { Err(mut out) => {
println!("{:?}", file_to_str(&mut out)); println!("{:?}", file_to_str(&mut out));
None None
} }
}; }
} }
} }

View file

@ -16,9 +16,9 @@ pub struct EvaluationFilterWorker {
impl EvaluationFilterWorker { impl EvaluationFilterWorker {
pub fn new(acl: acl::ACL) -> EvaluationFilterWorker { pub fn new(acl: acl::ACL) -> EvaluationFilterWorker {
return EvaluationFilterWorker { EvaluationFilterWorker {
acl, acl,
}; }
} }
} }
@ -31,7 +31,7 @@ impl worker::SimpleWorker for EvaluationFilterWorker {
_: &BasicProperties, _: &BasicProperties,
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
return match serde_json::from_slice(body) { match serde_json::from_slice(body) {
Ok(e) => Ok(e), Ok(e) => Ok(e),
Err(e) => { Err(e) => {
Err(format!( Err(format!(
@ -40,7 +40,7 @@ impl worker::SimpleWorker for EvaluationFilterWorker {
String::from_utf8(body.clone()) String::from_utf8(body.clone())
)) ))
} }
}; }
} }
fn consumer(&mut self, job: &ghevent::PullRequestEvent) -> worker::Actions { fn consumer(&mut self, job: &ghevent::PullRequestEvent) -> worker::Actions {

View file

@ -21,10 +21,10 @@ pub struct GitHubCommentWorker {
impl GitHubCommentWorker { impl GitHubCommentWorker {
pub fn new(acl: acl::ACL, github: hubcaps::Github) -> GitHubCommentWorker { pub fn new(acl: acl::ACL, github: hubcaps::Github) -> GitHubCommentWorker {
return GitHubCommentWorker { GitHubCommentWorker {
acl, acl,
github, github,
}; }
} }
} }
@ -37,7 +37,7 @@ impl worker::SimpleWorker for GitHubCommentWorker {
_: &BasicProperties, _: &BasicProperties,
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
return match serde_json::from_slice(body) { match serde_json::from_slice(body) {
Ok(e) => Ok(e), Ok(e) => Ok(e),
Err(e) => { Err(e) => {
println!( println!(
@ -46,7 +46,7 @@ impl worker::SimpleWorker for GitHubCommentWorker {
); );
panic!("{:?}", e); panic!("{:?}", e);
} }
}; }
} }
fn consumer(&mut self, job: &ghevent::IssueComment) -> worker::Actions { fn consumer(&mut self, job: &ghevent::IssueComment) -> worker::Actions {
@ -146,6 +146,6 @@ impl worker::SimpleWorker for GitHubCommentWorker {
} }
response.push(worker::Action::Ack); response.push(worker::Action::Ack);
return response; response
} }
} }

View file

@ -17,7 +17,7 @@ pub struct GitHubCommentPoster {
impl GitHubCommentPoster { impl GitHubCommentPoster {
pub fn new(github: hubcaps::Github) -> GitHubCommentPoster { pub fn new(github: hubcaps::Github) -> GitHubCommentPoster {
return GitHubCommentPoster { github }; GitHubCommentPoster { github }
} }
} }
@ -30,7 +30,7 @@ impl worker::SimpleWorker for GitHubCommentPoster {
_: &BasicProperties, _: &BasicProperties,
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
return match serde_json::from_slice(body) { match serde_json::from_slice(body) {
Ok(e) => Ok(e), Ok(e) => Ok(e),
Err(e) => { Err(e) => {
Err(format!( Err(format!(
@ -39,7 +39,7 @@ impl worker::SimpleWorker for GitHubCommentPoster {
e e
)) ))
} }
}; }
} }
fn consumer(&mut self, job: &BuildResult) -> worker::Actions { fn consumer(&mut self, job: &BuildResult) -> worker::Actions {
@ -70,7 +70,7 @@ impl worker::SimpleWorker for GitHubCommentPoster {
} }
} }
return vec![worker::Action::Ack]; vec![worker::Action::Ack]
} }
} }
@ -225,7 +225,7 @@ fn list_segment(name: &str, things: Vec<String>) -> Vec<String> {
reply.push("".to_owned()); reply.push("".to_owned());
} }
return reply; reply
} }
fn partial_log_segment(output: &Vec<String>) -> Vec<String> { fn partial_log_segment(output: &Vec<String>) -> Vec<String> {
@ -240,7 +240,7 @@ fn partial_log_segment(output: &Vec<String>) -> Vec<String> {
reply.push("```".to_owned()); reply.push("```".to_owned());
reply.push("</p></details>".to_owned()); reply.push("</p></details>".to_owned());
return reply; reply
} }
#[cfg(test)] #[cfg(test)]

View file

@ -52,18 +52,18 @@ fn validate_path_segment(segment: &PathBuf) -> Result<(), String> {
} }
}) })
{ {
return Ok(()); Ok(())
} else { } else {
return Err(String::from("Path contained invalid components")); Err(String::from("Path contained invalid components"))
} }
} }
impl LogMessageCollector { impl LogMessageCollector {
pub fn new(log_root: PathBuf, max_open: usize) -> LogMessageCollector { pub fn new(log_root: PathBuf, max_open: usize) -> LogMessageCollector {
return LogMessageCollector { LogMessageCollector {
handles: LruCache::new(max_open), handles: LruCache::new(max_open),
log_root, log_root,
}; }
} }
pub fn write_metadata(&mut self, from: &LogFrom, data: &BuildLogStart) -> Result<(), String>{ pub fn write_metadata(&mut self, from: &LogFrom, data: &BuildLogStart) -> Result<(), String>{
@ -105,20 +105,20 @@ impl LogMessageCollector {
pub fn handle_for(&mut self, from: &LogFrom) -> Result<&mut LineWriter, String> { pub fn handle_for(&mut self, from: &LogFrom) -> Result<&mut LineWriter, String> {
if self.handles.contains_key(&from) { if self.handles.contains_key(&from) {
return Ok(self.handles.get_mut(&from).expect( Ok(self.handles.get_mut(&from).expect(
"handles just contained the key", "handles just contained the key",
)); ))
} else { } else {
let logpath = self.path_for_log(&from)?; let logpath = self.path_for_log(&from)?;
let fp = self.open_file(logpath)?; let fp = self.open_file(logpath)?;
let writer = LineWriter::new(fp); let writer = LineWriter::new(fp);
self.handles.insert(from.clone(), writer); self.handles.insert(from.clone(), writer);
if let Some(handle) = self.handles.get_mut(&from) { if let Some(handle) = self.handles.get_mut(&from) {
return Ok(handle); Ok(handle)
} else { } else {
return Err(String::from( Err(String::from(
"A just-inserted value should already be there", "A just-inserted value should already be there",
)); ))
} }
} }
} }
@ -126,13 +126,13 @@ impl LogMessageCollector {
fn path_for_metadata(&self, from: &LogFrom) -> Result<PathBuf, String> { fn path_for_metadata(&self, from: &LogFrom) -> Result<PathBuf, String> {
let mut path = self.path_for_log(from)?; let mut path = self.path_for_log(from)?;
path.set_extension("metadata.json"); path.set_extension("metadata.json");
return Ok(path); Ok(path)
} }
fn path_for_result(&self, from: &LogFrom) -> Result<PathBuf, String> { fn path_for_result(&self, from: &LogFrom) -> Result<PathBuf, String> {
let mut path = self.path_for_log(from)?; let mut path = self.path_for_log(from)?;
path.set_extension("result.json"); path.set_extension("result.json");
return Ok(path); Ok(path)
} }
fn path_for_log(&self, from: &LogFrom) -> Result<PathBuf, String> { fn path_for_log(&self, from: &LogFrom) -> Result<PathBuf, String> {
@ -147,13 +147,13 @@ impl LogMessageCollector {
location.push(attempt_id); location.push(attempt_id);
if location.starts_with(&self.log_root) { if location.starts_with(&self.log_root) {
return Ok(location); Ok(location)
} else { } else {
return Err(format!( Err(format!(
"Calculating the log location for {:?} resulted in an invalid path {:?}", "Calculating the log location for {:?} resulted in an invalid path {:?}",
from, from,
location location
)); ))
} }
} }
@ -212,13 +212,13 @@ impl worker::SimpleWorker for LogMessageCollector {
} }
} }
return Ok(LogMessage { Ok(LogMessage {
from: LogFrom { from: LogFrom {
routing_key: deliver.routing_key.clone(), routing_key: deliver.routing_key.clone(),
attempt_id, attempt_id,
}, },
message, message,
}); })
} }
fn consumer(&mut self, job: &LogMessage) -> worker::Actions { fn consumer(&mut self, job: &LogMessage) -> worker::Actions {
@ -237,7 +237,7 @@ impl worker::SimpleWorker for LogMessageCollector {
}, },
} }
return vec![worker::Action::Ack]; vec![worker::Action::Ack]
} }
} }

View file

@ -45,7 +45,7 @@ impl<E: stats::SysEvents> MassRebuildWorker<E> {
events: E, events: E,
tag_paths: HashMap<String, Vec<String>>, tag_paths: HashMap<String, Vec<String>>,
) -> MassRebuildWorker<E> { ) -> MassRebuildWorker<E> {
return MassRebuildWorker { MassRebuildWorker {
cloner, cloner,
nix: nix.without_limited_supported_systems(), nix: nix.without_limited_supported_systems(),
github, github,
@ -53,11 +53,11 @@ impl<E: stats::SysEvents> MassRebuildWorker<E> {
identity, identity,
events, events,
tag_paths, tag_paths,
}; }
} }
fn actions(&self) -> massrebuildjob::Actions { fn actions(&self) -> massrebuildjob::Actions {
return massrebuildjob::Actions {}; massrebuildjob::Actions {}
} }
fn tag_from_title(&self, issue: &hubcaps::issues::IssueRef) { fn tag_from_title(&self, issue: &hubcaps::issues::IssueRef) {
@ -102,7 +102,7 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for MassRebuildWorker<E
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
self.events.notify(Event::JobReceived); self.events.notify(Event::JobReceived);
return match massrebuildjob::from(body) { match massrebuildjob::from(body) {
Ok(e) => { Ok(e) => {
self.events.notify(Event::JobDecodeSuccess); self.events.notify(Event::JobDecodeSuccess);
Ok(e) Ok(e)
@ -116,7 +116,7 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for MassRebuildWorker<E
); );
Err("Failed to decode message".to_owned()) Err("Failed to decode message".to_owned())
} }
}; }
} }
fn consumer(&mut self, job: &massrebuildjob::MassRebuildJob) -> worker::Actions { fn consumer(&mut self, job: &massrebuildjob::MassRebuildJob) -> worker::Actions {
@ -450,9 +450,9 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for MassRebuildWorker<E
status.set(state.clone()); status.set(state.clone());
if state == hubcaps::statuses::State::Success { if state == hubcaps::statuses::State::Success {
return Ok(()); Ok(())
} else { } else {
return Err(()); Err(())
} }
}) })
.all(|status| status == Ok(())); .all(|status| status == Ok(()));
@ -585,7 +585,7 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for MassRebuildWorker<E
self.events.notify(Event::TaskEvaluationCheckComplete); self.events.notify(Event::TaskEvaluationCheckComplete);
return self.actions().done(&job, response); self.actions().done(&job, response)
} }
} }
@ -606,7 +606,7 @@ fn make_gist<'a>(
}, },
); );
return Some( Some(
gists gists
.create(&hubcaps::gists::GistOptions { .create(&hubcaps::gists::GistOptions {
description, description,
@ -615,7 +615,7 @@ fn make_gist<'a>(
}) })
.expect("Failed to create gist!") .expect("Failed to create gist!")
.html_url, .html_url,
); )
} }
pub fn update_labels(issue: &hubcaps::issues::IssueRef, add: Vec<String>, remove: Vec<String>) { pub fn update_labels(issue: &hubcaps::issues::IssueRef, add: Vec<String>, remove: Vec<String>) {
@ -735,5 +735,5 @@ fn indicates_wip(text: &str) -> bool {
return true; return true;
} }
return false; false
} }

View file

@ -29,7 +29,7 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for StatCollectorWorker
_: &BasicProperties, _: &BasicProperties,
body: &Vec<u8>, body: &Vec<u8>,
) -> Result<Self::J, String> { ) -> Result<Self::J, String> {
return match serde_json::from_slice(body) { match serde_json::from_slice(body) {
Ok(e) => Ok(e), Ok(e) => Ok(e),
Err(_) => { Err(_) => {
let mut modified_body: Vec<u8> = vec!["\"".as_bytes()[0]]; let mut modified_body: Vec<u8> = vec!["\"".as_bytes()[0]];
@ -55,7 +55,7 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for StatCollectorWorker
} }
} }
} }
}; }
} }
fn consumer(&mut self, job: &stats::EventMessage) -> worker::Actions { fn consumer(&mut self, job: &stats::EventMessage) -> worker::Actions {
@ -65,6 +65,6 @@ impl<E: stats::SysEvents + 'static> worker::SimpleWorker for StatCollectorWorker
self.collector.record(sender.clone(), event.clone()); self.collector.record(sender.clone(), event.clone());
} }
return vec![worker::Action::Ack]; vec![worker::Action::Ack]
} }
} }

View file

@ -17,7 +17,7 @@ impl TestScratch {
TestScratch::create_dir(&scratch); TestScratch::create_dir(&scratch);
return scratch; scratch
} }
pub fn new_file(ident: &str) -> TestScratch { pub fn new_file(ident: &str) -> TestScratch {
@ -29,7 +29,7 @@ impl TestScratch {
}; };
TestScratch::create_dir(&scratch); TestScratch::create_dir(&scratch);
return scratch; scratch
} }
fn create_dir(path: &TestScratch) { fn create_dir(path: &TestScratch) {

View file

@ -44,14 +44,14 @@ where
..Default::default() ..Default::default()
}; };
return Action::Publish(QueueMsg { Action::Publish(QueueMsg {
exchange, exchange,
routing_key, routing_key,
mandatory: false, mandatory: false,
immediate: false, immediate: false,
properties: Some(props), properties: Some(props),
content: serde_json::to_string(&msg).unwrap().into_bytes(), content: serde_json::to_string(&msg).unwrap().into_bytes(),
}); })
} }
pub trait SimpleWorker: Send + 'static { pub trait SimpleWorker: Send + 'static {
@ -68,7 +68,7 @@ pub trait SimpleWorker: Send + 'static {
} }
pub fn new<T: SimpleWorker>(worker: T) -> Worker<T> { pub fn new<T: SimpleWorker>(worker: T) -> Worker<T> {
return Worker { internal: worker }; Worker { internal: worker }
} }

View file

@ -22,7 +22,7 @@ impl LineWriter {
last_line: len, last_line: len,
}; };
return writer; writer
} }
fn load_buffer(file: &mut File) -> Vec<String> { fn load_buffer(file: &mut File) -> Vec<String> {