Codebase list libmediawiki-bot-perl / HEAD
HEAD

Tree @HEAD (Download .tar.gz)

   1
   2
   3
   4
   5
   6
   7
   8
   9
  10
  11
  12
  13
  14
  15
  16
  17
  18
  19
  20
  21
  22
  23
  24
  25
  26
  27
  28
  29
  30
  31
  32
  33
  34
  35
  36
  37
  38
  39
  40
  41
  42
  43
  44
  45
  46
  47
  48
  49
  50
  51
  52
  53
  54
  55
  56
  57
  58
  59
  60
  61
  62
  63
  64
  65
  66
  67
  68
  69
  70
  71
  72
  73
  74
  75
  76
  77
  78
  79
  80
  81
  82
  83
  84
  85
  86
  87
  88
  89
  90
  91
  92
  93
  94
  95
  96
  97
  98
  99
 100
 101
 102
 103
 104
 105
 106
 107
 108
 109
 110
 111
 112
 113
 114
 115
 116
 117
 118
 119
 120
 121
 122
 123
 124
 125
 126
 127
 128
 129
 130
 131
 132
 133
 134
 135
 136
 137
 138
 139
 140
 141
 142
 143
 144
 145
 146
 147
 148
 149
 150
 151
 152
 153
 154
 155
 156
 157
 158
 159
 160
 161
 162
 163
 164
 165
 166
 167
 168
 169
 170
 171
 172
 173
 174
 175
 176
 177
 178
 179
 180
 181
 182
 183
 184
 185
 186
 187
 188
 189
 190
 191
 192
 193
 194
 195
 196
 197
 198
 199
 200
 201
 202
 203
 204
 205
 206
 207
 208
 209
 210
 211
 212
 213
 214
 215
 216
 217
 218
 219
 220
 221
 222
 223
 224
 225
 226
 227
 228
 229
 230
 231
 232
 233
 234
 235
 236
 237
 238
 239
 240
 241
 242
 243
 244
 245
 246
 247
 248
 249
 250
 251
 252
 253
 254
 255
 256
 257
 258
 259
 260
 261
 262
 263
 264
 265
 266
 267
 268
 269
 270
 271
 272
 273
 274
 275
 276
 277
 278
 279
 280
 281
 282
 283
 284
 285
 286
 287
 288
 289
 290
 291
 292
 293
 294
 295
 296
 297
 298
 299
 300
 301
 302
 303
 304
 305
 306
 307
 308
 309
 310
 311
 312
 313
 314
 315
 316
 317
 318
 319
 320
 321
 322
 323
 324
 325
 326
 327
 328
 329
 330
 331
 332
 333
 334
 335
 336
 337
 338
 339
 340
 341
 342
 343
 344
 345
 346
 347
 348
 349
 350
 351
 352
 353
 354
 355
 356
 357
 358
 359
 360
 361
 362
 363
 364
 365
 366
 367
 368
 369
 370
 371
 372
 373
 374
 375
 376
 377
 378
 379
 380
 381
 382
 383
 384
 385
 386
 387
 388
 389
 390
 391
 392
 393
 394
 395
 396
 397
 398
 399
 400
 401
 402
 403
 404
 405
 406
 407
 408
 409
 410
 411
 412
 413
 414
 415
 416
 417
 418
 419
 420
 421
 422
 423
 424
 425
 426
 427
 428
 429
 430
 431
 432
 433
 434
 435
 436
 437
 438
 439
 440
 441
 442
 443
 444
 445
 446
 447
 448
 449
 450
 451
 452
 453
 454
 455
 456
 457
 458
 459
 460
 461
 462
 463
 464
 465
 466
 467
 468
 469
 470
 471
 472
 473
 474
 475
 476
 477
 478
 479
 480
 481
 482
 483
 484
 485
 486
 487
 488
 489
 490
 491
 492
 493
 494
 495
 496
 497
 498
 499
 500
 501
 502
 503
 504
 505
 506
 507
 508
 509
 510
 511
 512
 513
 514
 515
 516
 517
 518
 519
 520
 521
 522
 523
 524
 525
 526
 527
 528
 529
 530
 531
 532
 533
 534
 535
 536
 537
 538
 539
 540
 541
 542
 543
 544
 545
 546
 547
 548
 549
 550
 551
 552
 553
 554
 555
 556
 557
 558
 559
 560
 561
 562
 563
 564
 565
 566
 567
 568
 569
 570
 571
 572
 573
 574
 575
 576
 577
 578
 579
 580
 581
 582
 583
 584
 585
 586
 587
 588
 589
 590
 591
 592
 593
 594
 595
 596
 597
 598
 599
 600
 601
 602
 603
 604
 605
 606
 607
 608
 609
 610
 611
 612
 613
 614
 615
 616
 617
 618
 619
 620
 621
 622
 623
 624
 625
 626
 627
 628
 629
 630
 631
 632
 633
 634
 635
 636
 637
 638
 639
 640
 641
 642
 643
 644
 645
 646
 647
 648
 649
 650
 651
 652
 653
 654
 655
 656
 657
 658
 659
 660
 661
 662
 663
 664
 665
 666
 667
 668
 669
 670
 671
 672
 673
 674
 675
 676
 677
 678
 679
 680
 681
 682
 683
 684
 685
 686
 687
 688
 689
 690
 691
 692
 693
 694
 695
 696
 697
 698
 699
 700
 701
 702
 703
 704
 705
 706
 707
 708
 709
 710
 711
 712
 713
 714
 715
 716
 717
 718
 719
 720
 721
 722
 723
 724
 725
 726
 727
 728
 729
 730
 731
 732
 733
 734
 735
 736
 737
 738
 739
 740
 741
 742
 743
 744
 745
 746
 747
 748
 749
 750
 751
 752
 753
 754
 755
 756
 757
 758
 759
 760
 761
 762
 763
 764
 765
 766
 767
 768
 769
 770
 771
 772
 773
 774
 775
 776
 777
 778
 779
 780
 781
 782
 783
 784
 785
 786
 787
 788
 789
 790
 791
 792
 793
 794
 795
 796
 797
 798
 799
 800
 801
 802
 803
 804
 805
 806
 807
 808
 809
 810
 811
 812
 813
 814
 815
 816
 817
 818
 819
 820
 821
 822
 823
 824
 825
 826
 827
 828
 829
 830
 831
 832
 833
 834
 835
 836
 837
 838
 839
 840
 841
 842
 843
 844
 845
 846
 847
 848
 849
 850
 851
 852
 853
 854
 855
 856
 857
 858
 859
 860
 861
 862
 863
 864
 865
 866
 867
 868
 869
 870
 871
 872
 873
 874
 875
 876
 877
 878
 879
 880
 881
 882
 883
 884
 885
 886
 887
 888
 889
 890
 891
 892
 893
 894
 895
 896
 897
 898
 899
 900
 901
 902
 903
 904
 905
 906
 907
 908
 909
 910
 911
 912
 913
 914
 915
 916
 917
 918
 919
 920
 921
 922
 923
 924
 925
 926
 927
 928
 929
 930
 931
 932
 933
 934
 935
 936
 937
 938
 939
 940
 941
 942
 943
 944
 945
 946
 947
 948
 949
 950
 951
 952
 953
 954
 955
 956
 957
 958
 959
 960
 961
 962
 963
 964
 965
 966
 967
 968
 969
 970
 971
 972
 973
 974
 975
 976
 977
 978
 979
 980
 981
 982
 983
 984
 985
 986
 987
 988
 989
 990
 991
 992
 993
 994
 995
 996
 997
 998
 999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
SYNOPSIS

        use MediaWiki::Bot qw(:constants);
    
        my $bot = MediaWiki::Bot->new({
            assert      => 'bot',
            host        => 'de.wikimedia.org',
            login_data  => { username => "Mike's bot account", password => "password" },
        });
    
        my $revid = $bot->get_last("User:Mike.lifeguard/sandbox", "Mike.lifeguard");
        print "Reverting to $revid\n" if defined($revid);
        $bot->revert('User:Mike.lifeguard', $revid, 'rvv');

DESCRIPTION

    MediaWiki::Bot is a framework that can be used to write bots which
    interface with the MediaWiki API (http://en.wikipedia.org/w/api.php).

METHODS

 new

        my $bot = MediaWiki::Bot({
            host     => 'en.wikipedia.org',
            operator => 'Mike.lifeguard',
        });

    Calling MediaWiki::Bot->new() will create a new MediaWiki::Bot object.
    The only parameter is a hashref with keys:

      * agent sets a custom useragent. It is recommended to use operator
      instead, which is all we need to do the right thing for you. If you
      really want to do it yourself, see
      https://meta.wikimedia.org/wiki/User-agent_policy for guidance on
      what information must be included.

      * assert sets a parameter for the AssertEdit extension (commonly
      'bot')

      Refer to http://mediawiki.org/wiki/Extension:AssertEdit.

      * operator allows the bot to send you a message when it fails an
      assert. This is also the recommended way to customize the user agent
      string, which is required by the Wikimedia Foundation. A warning will
      be emitted if you omit this.

      * maxlag allows you to set the maxlag parameter (default is the
      recommended 5s).

      Please refer to the MediaWiki documentation prior to changing this
      from the default.

      * protocol allows you to specify 'http' or 'https' (default is
      'http')

      * host sets the domain name of the wiki to connect to

      * path sets the path to api.php (with no leading or trailing slash)

      * login_data is a hashref of credentials to pass to "login".

      * debug - whether to provide debug output.

      1 provides only error messages; 2 provides further detail on internal
      operations.

    For example:

        my $bot = MediaWiki::Bot->new({
            assert      => 'bot',
            protocol    => 'https',
            host        => 'en.wikimedia.org',
            agent       => sprintf(
                'PerlWikiBot/%s (https://metacpan.org/MediaWiki::Bot; User:Mike.lifeguard)',
                MediaWiki::Bot->VERSION
            ),
            login_data  => { username => "Mike's bot account", password => "password" },
        });

    For backward compatibility, you can specify up to three parameters:

        my $bot = MediaWiki::Bot->new('My custom useragent string', $assert, $operator);

    This form is deprecated will never do auto-login or autoconfiguration,
    and emits deprecation warnings.

    For further reading:

      * MediaWiki::Bot wiki
      <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki>

      * <Installing MediaWiki::Bot
      <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Install>>

      * Creating a new bot
      <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Creating-a-new-bot>

      * Setting the wiki
      <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Setting-the-wiki>

      * Where is api.php
      <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Where-is-api.php>

 set_wiki

    Set what wiki to use. The parameter is a hashref with keys:

      * host - the domain name

      * path - the part of the path before api.php (usually 'w')

      * protocol is either 'http' or 'https'.

    If you don't set any parameter, it's previous value is used. If it has
    never been set, the default settings are 'http', 'en.wikipedia.org' and
    'w'.

    For example:

        $bot->set_wiki({
            protocol    => 'https',
            host        => 'secure.wikimedia.org',
            path        => 'wikipedia/meta/w',
        });

    For backward compatibility, you can specify up to two parameters:

        $bot->set_wiki($host, $path);

    This form is deprecated, and will emit deprecation warnings.

 login

    This method takes a hashref with keys username and password at a
    minimum. See "Single User Login" and "Basic authentication" for
    additional options.

    Logs the use $username in, optionally using $password. First, an
    attempt will be made to use cookies to log in. If this fails, an
    attempt will be made to use the password provided to log in, if any. If
    the login was successful, returns true; false otherwise.

        $bot->login({
            username => $username,
            password => $password,
        }) or die "Login failed";

    Once logged in, attempt to do some simple auto-configuration. At
    present, this consists of:

      * Warning if the account doesn't have the bot flag, and isn't a sysop
      account.

      * Setting an appropriate default assert.

    You can skip this autoconfiguration by passing autoconfig => 0

    For backward compatibility, you can call this as

        $bot->login($username, $password);

    This form is deprecated, and will emit deprecation warnings. It will
    never do autoconfiguration or SUL login.

  Single User Login

    On WMF wikis, do_sul specifies whether to log in on all projects. The
    default is false. But even when false, you still get a CentralAuth
    cookie for, and are thus logged in on, all languages of a given domain
    (*.wikipedia.org, for example). When set, a login is done on each WMF
    domain so you are logged in on all ~800 content wikis. Since
    *.wikimedia.org is not possible, we explicitly include meta, commons,
    incubator, and wikispecies.

  Basic authentication

    If you need to supply basic auth credentials, pass a hashref of data as
    described by LWP::UserAgent:

        $bot->login({
            username    => $username,
            password    => $password,
            basic_auth  => {    netloc  => "private.wiki.com:80",
                                realm   => "Authentication Realm",
                                uname   => "Basic auth username",
                                pass    => "password",
                            }
        }) or die "Couldn't log in";

  Bot passwords

    MediaWiki::Bot doesn't yet support the more complicated (but more
    secure) oAuth login flow for bots. Instead, we support a simpler "bot
    password", which is a generated password connected to a
    (possibly-reduced) set of on-wiki privileges, and IP ranges from which
    it can be used.

    To create one, visit Special:BotPasswords on the wiki. Enter a label
    for the password, then select the privileges you want to use with that
    password. This set should be as restricted as possible; most bots only
    edit existing pages. Keeping the set of privileges as restricted as
    possible limits the possible damage if the password were ever
    compromised.

    Submit the form, and you'll be given a new "username" that looks like
    "AccountUsername@bot_password_label", and a generated bot password. To
    log in, provide those to MediaWiki::Bot verbatim.

    References: API:Login <https://www.mediawiki.org/wiki/API:Login>,
    Logging in
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Logging-in>

 logout

        $bot->logout();

    The logout method logs the bot out of the wiki. This invalidates all
    login cookies.

    References: API:Logging out <https://www.mediawiki.org/wiki/API:Logout>

 edit

        my $text = $bot->get_text('My page');
        $text .= "\n\n* More text\n";
        $bot->edit({
            page    => 'My page',
            text    => $text,
            summary => 'Adding new content',
            section => 'new',
        });

    This method edits a wiki page, and takes a hashref of data with keys:

      * page - the page title to edit

      * text - the page text to write

      * summary - an edit summary

      * minor - whether to mark the edit as minor or not (boolean)

      * bot - whether to mark the edit as a bot edit (boolean)

      * assertion - usually 'bot', but see
      http://mediawiki.org/wiki/Extension:AssertEdit.

      * section - edit a single section (identified by number) instead of
      the whole page

    An MD5 hash is sent to guard against data corruption while in transit.

    You can also call this as:

        $bot->edit($page, $text, $summary, $is_minor, $assert, $markasbot);

    This form is deprecated, and will emit deprecation warnings.

  CAPTCHAs

    If a CAPTCHA <https://en.wikipedia.org/wiki/CAPTCHA> is encountered,
    the call to edit will return false, with the error code set to
    ERR_CAPTCHA and the details informing you that solving a CAPTCHA is
    required for this action. The information you need to actually solve
    the captcha (for example the URL for the image) is given in
    $bot->{error}->{captcha} as a hash reference. You will want to grab the
    keys 'url' (a relative URL to the image) and 'id' (the ID of the
    CAPTCHA). Once you have solved the CAPTCHA (presumably by interacting
    with a human), retry the edit, adding captcha_id and captcha_solution
    parameters:

        my $edit = {page => 'Main Page', text => 'got your nose'};
        my $edit_status = $bot->edit($edit);
        if (not $edit_status) {
            if ($bot->{error}->{code} == ERR_CAPTCHA) {
                my @captcha_uri = split /\Q?/, $bot->{error}{captcha}{url}, 2;
                my $image = URI->new(sprintf '%s://%s%s?%s' =>
                    $bot->{protocol}, $bot->{host}, $captcha_uri[0], $captcha_uri[1],
                );
    
                require Term::ReadLine;
                my $term = Term::ReadLine->new('Solve the captcha');
                $term->ornaments(0);
                my $answer = $term->readline("Please solve $image and type the answer: ");
    
                # Add new CAPTCHA params to the edit we're attempting
                $edit->{captcha_id} = $bot->{error}->{captcha}->{id};
                $edit->{captcha_solution} = $answer;
                $status = $bot->edit($edit);
            }
        }

    References: Editing pages
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Editing-pages>,
    API:Edit <https://www.mediawiki.org/wiki/API:Edit>, API:Tokens
    <https://www.mediawiki.org/wiki/API:Tokens>

 move

        $bot->move($from_title, $to_title, $reason, $options_hashref);

    This moves a wiki page.

    If you wish to specify more options (like whether to suppress creation
    of a redirect), use $options_hashref, which has keys:

      * movetalk specifies whether to attempt to the talk page.

      * noredirect specifies whether to suppress creation of a redirect.

      * movesubpages specifies whether to move subpages, if applicable.

      * watch and unwatch add or remove the page and the redirect from your
      watchlist.

      * ignorewarnings ignores warnings.

        my @pages = ("Humor", "Rumor");
        foreach my $page (@pages) {
            my $to = $page;
            $to =~ s/or$/our/;
            $bot->move($page, $to, "silly 'merricans");
        }

    References: API:Move <https://www.mediawiki.org/wiki/API:Move>

 get_history

        my @hist = $bot->get_history($title);
        my @hist = $bot->get_history($title, $additional_params);

    Returns an array containing the history of the specified page $title.

    The optional hash ref $additional_params can be used to tune the query
    by API parameters, such as 'rvlimit' to return only 'rvlimit' number of
    revisions (default is as many as possible, but may be limited per
    query) or 'rvdir' to set the chronological direction.

    Example:

        my @hist = $bot->get_history('Main Page', {'rvlimit' => 10, 'rvdir' => 'older'})

    The array returned contains hashrefs with keys: revid, user, comment,
    minor, timestamp_date, and timestamp_time.

    For backward compatibility, you can specify up to four parameters:

        my @hist = $bot->get_history($title, $limit, $revid, $direction);

    References: Getting page history
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Getting-page-history>,
    API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 get_history_step_by_step

        my @hist = $bot->get_history_step_by_step($title);
        my @hist = $bot->get_history_step_by_step($title, $additional_params);

    Same as get_history(), but does not return the full history at once,
    but let's you loop through it.

    The optional call-by-reference hash ref $additional_params can be used
    to loop through a page's full history by using the 'continue' param
    returned by the API.

    Example:

        my $ready;
        my $filter_params = {};
        while(!$ready){
            my @hist = $bot->get_history_step_by_step($page, $filter_params);
            if(@hist == 0 || !defined($filter_params->{'continue'})){
                $ready = 1;
            }
            # do something with @hist
        }

    References: Getting page history
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Getting-page-history>,
    API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 get_text

    Returns the wikitext of the specified $page_title. The first parameter
    $page_title is the only required one.

    The second parameter is a hashref with the following independent
    optional keys:

      * rvstartid - if defined, this function returns the text of that
      revision, otherwise the newest revision will be used.

      * rvsection - if defined, returns the text of that section. Otherwise
      the whole page text will be returned.

      * pageid - this is an output parameter and can be used to fetch the
      id of a page without the need of calling "get_id" additionally. Note
      that the value of this param is ignored and it will be overwritten by
      this function.

      * rv... - any param starting with 'rv' will be forwarded to the api
      call.

    A blank page will return wikitext of "" (which evaluates to false in
    Perl, but is defined); a nonexistent page will return undef (which also
    evaluates to false in Perl, but is obviously undefined). You can
    distinguish between blank and nonexistent pages by using defined:

        # simple example
        my $wikitext = $bot->get_text('Page title');
        print "Wikitext: $wikitext\n" if defined $wikitext;
    
        # advanced example
        my $options = {'revid'=>123456, 'section_number'=>2};
        $wikitext = $bot->get_text('Page title', $options);
        die "error, see API error message\n" unless defined $options->{'pageid'};
        warn "page doesn't exist\n" if $options->{'pageid'} == MediaWiki::Bot::PAGE_NONEXISTENT;
        print "Wikitext: $wikitext\n" if defined $wikitext;

    References: Fetching page text
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Fetching-page-text>,
    API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

    For backward-compatibility the params revid and section_number may also
    be given as scalar parameters:

        my $wikitext = $bot->get_text('Page title', 123456, 2);
        print "Wikitext: $wikitext\n" if defined $wikitext;

 get_id

    Returns the id of the specified $page_title. Returns undef if page does
    not exist.

        my $pageid = $bot->get_id("Main Page");
        die "Page doesn't exist\n" if !defined($pageid);

    Revisions: API:Properties#info
    <https://www.mediawiki.org/wiki/API:Properties#info_.2F_in>

 get_pages

    Returns the text of the specified pages in a hashref. Content of undef
    means page does not exist. Also handles redirects or article names that
    use namespace aliases.

        my @pages = ('Page 1', 'Page 2', 'Page 3');
        my $thing = $bot->get_pages(\@pages);
        foreach my $page (keys %$thing) {
            my $text = $thing->{$page};
            print "$text\n" if defined($text);
        }

    References: Fetching page text
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Fetching-page-text>,
    API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 get_image

        $buffer = $bot->get_image('File:Foo.jpg', { width=>256, height=>256 });

    Download an image from a wiki. This is derived from a similar function
    in MediaWiki::API. This one allows the image to be scaled down by
    passing a hashref with height & width parameters.

    It returns raw data in the original format. You may simply spew it to a
    file, or process it directly with a library such as Imager.

        use File::Slurp qw(write_file);
        my $img_data = $bot->get_image('File:Foo.jpg');
        write_file( 'Foo.jpg', {binmode => ':raw'}, \$img_data );

    Images are scaled proportionally. (height/width) will remain constant,
    except for rounding errors.

    Height and width parameters describe the maximum dimensions. A 400x200
    image will never be scaled to greater dimensions. You can scale it
    yourself; having the wiki do it is just lazy & selfish.

    References: API:Properties#imageinfo
    <https://www.mediawiki.org/wiki/API:Properties#imageinfo_.2F_ii>

 revert

    Reverts the specified $page_title to $revid, with an edit summary of
    $summary. A default edit summary will be used if $summary is omitted.

        my $revid = $bot->get_last("User:Mike.lifeguard/sandbox", "Mike.lifeguard");
        print "Reverting to $revid\n" if defined($revid);
        $bot->revert('User:Mike.lifeguard', $revid, 'rvv');

    References: API:Edit <https://www.mediawiki.org/wiki/API:Edit>

 undo

        $bot->undo($title, $revid, $summary, $after);

    Reverts the specified $revid, with an edit summary of $summary, using
    the undo function. To undo all revisions from $revid up to but not
    including this one, set $after to another revid. If not set, just undo
    the one revision ($revid).

    References: API:Edit <https://www.mediawiki.org/wiki/API:Edit>

 get_last

    Returns the revid of the last revision to $page not made by $user.
    undef is returned if no result was found, as would be the case if the
    page is deleted.

        my $revid = $bot->get_last('User:Mike.lifeguard/sandbox', 'Mike.lifeguard');
        if defined($revid) {
            print "Reverting to $revid\n";
            $bot->revert('User:Mike.lifeguard', $revid, 'rvv');
        }

    References: API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 update_rc

    This method is deprecated, and will emit deprecation warnings. Replace
    calls to update_rc() with calls to the newer recentchanges(), which
    returns all available data, including rcid.

    Returns an array containing the $limit most recent changes to the
    wiki's main namespace. The array contains hashrefs with keys title,
    revid, old_revid, and timestamp.

        my @rc = $bot->update_rc(5);
        foreach my $hashref (@rc) {
            my $title = $hash->{'title'};
            print "$title\n";
        }

    The "Options hashref" is also available:

        # Use a callback for incremental processing:
        my $options = { hook => \&mysub, };
        $bot->update_rc($options);
        sub mysub {
            my ($res) = @_;
            foreach my $hashref (@$res) {
                my $page = $hashref->{'title'};
                print "$page\n";
            }
        }

 recentchanges($wiki_hashref, $options_hashref)

    Returns an array of hashrefs containing recentchanges data.

    The first parameter is a hashref with the following keys:

      * ns - the namespace number, or an arrayref of numbers to specify
      several; default is the main namespace

      * limit - the number of rows to fetch; default is 50

      * user - only list changes by this user

      * show - itself a hashref where the key is a category and the value
      is a boolean. If true, the category will be included; if false,
      excluded. The categories are kinds of edits: minor, bot, anon,
      redirect, patrolled. See "rcshow" at
      http://www.mediawiki.org/wiki/API:Recentchanges#Parameters.

    An "Options hashref" can be used as the second parameter:

        my @rc = $bot->recentchanges({ ns => 4, limit => 100 });
        foreach my $hashref (@rc) {
            print $hashref->{title} . "\n";
        }
    
        # Or, use a callback for incremental processing:
        $bot->recentchanges({ ns => [0,1], limit => 500 }, { hook => \&mysub });
        sub mysub {
            my ($res) = @_;
            foreach my $hashref (@$res) {
                my $page = $hashref->{title};
                print "$page\n";
            }
        }

    The hashref returned might contain the following keys:

      * ns - the namespace number

      * revid

      * old_revid

      * timestamp

      * rcid - can be used with "patrol"

      * pageid

      * type - one of edit, new, log (there may be others)

      * title

    For backwards compatibility, the previous method signature is still
    supported:

        $bot->recentchanges($ns, $limit, $options_hashref);

    References: API:Recentchanges
    <https://www.mediawiki.org/wiki/API:Recentchanges>

 what_links_here

    Returns an array containing a list of all pages linking to $page.

    Additional optional parameters are:

      * One of: all (default), redirects, or nonredirects.

      * A namespace number to search (pass an arrayref to search in
      multiple namespaces)

      * An "Options hashref".

    A typical query:

        my @links = $bot->what_links_here("Meta:Sandbox",
            undef, 1,
            { hook=>\&mysub }
        );
        sub mysub{
            my ($res) = @_;
            foreach my $hash (@$res) {
                my $title = $hash->{'title'};
                my $is_redir = $hash->{'redirect'};
                print "Redirect: $title\n" if $is_redir;
                print "Page: $title\n" unless $is_redir;
            }
        }

    Transclusions are no longer handled by what_links_here() - use
    "list_transclusions" instead.

    References: Listing incoming links
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Listing-incoming-links>,
    API:Backlinks <https://www.mediawiki.org/wiki/API:Backlinks>

 list_transclusions

    Returns an array containing a list of all pages transcluding $page.

    Other parameters are:

      * One of: all (default), redirects, or nonredirects

      * A namespace number to search (pass an arrayref to search in
      multiple namespaces).

      * $options_hashref as described by MediaWiki::API:

      Set max to limit the number of queries performed.

      Set hook to a subroutine reference to use a callback hook for
      incremental processing.

      Refer to the section on "linksearch" for examples.

    A typical query:

        $bot->list_transclusions("Template:Tlx", undef, 4, {hook => \&mysub});
        sub mysub{
            my ($res) = @_;
            foreach my $hash (@$res) {
                my $title = $hash->{'title'};
                my $is_redir = $hash->{'redirect'};
                print "Redirect: $title\n" if $is_redir;
                print "Page: $title\n" unless $is_redir;
            }
        }

    References: Listing transclusions
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Listing-transclusions>
    API:Embeddedin <https://www.mediawiki.org/wiki/API:Embeddedin>

 get_pages_in_category

    Returns an array containing the names of all pages in the specified
    category (include the Category: prefix). Does not recurse into
    sub-categories.

        my @pages = $bot->get_pages_in_category('Category:People on stamps of Gabon');
        print "The pages in Category:People on stamps of Gabon are:\n@pages\n";

    The options hashref is as described in "Options hashref". Use { max =>
    0 } to get all results.

    References: Listing category contents
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Listing-category-contents>,
    API:Categorymembers
    <https://www.mediawiki.org/wiki/API:Categorymembers>

 get_all_pages_in_category

        my @pages = $bot->get_all_pages_in_category($category, $options_hashref);

    Returns an array containing the names of all pages in the specified
    category (include the Category: prefix), including sub-categories. The
    $options_hashref is described fully in "Options hashref".

    References: Listing category contents
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Listing-category-contents>,
    API:Categorymembers
    <https://www.mediawiki.org/wiki/API:Categorymembers>

 get_all_categories

    Returns an array containing the names of all categories.

        my @categories = $bot->get_all_categories();
        print "The categories are:\n@categories\n";

    Use { max => 0 } to get all results. The default number of categories
    returned is 10, the maximum allowed is 500.

    References: API:Allcategories
    <https://www.mediawiki.org/wiki/API:Allcategories>

 linksearch

    Runs a linksearch on the specified $link and returns an array
    containing anonymous hashes with keys 'url' for the outbound URL, and
    'title' for the page the link is on.

    Additional parameters are:

      * A namespace number to search (pass an arrayref to search in
      multiple namespaces).

      * You can search by $protocol (http is default).

      * $options_hashref is fully documented in "Options hashref":

      Set max in $options to get more than one query's worth of results:

          my $options = { max => 10, }; # I only want some results
          my @links = $bot->linksearch("slashdot.org", 1, undef, $options);
          foreach my $hash (@links) {
              my $url = $hash->{'url'};
              my $page = $hash->{'title'};
              print "$page: $url\n";
          }

      Set hook to a subroutine reference to use a callback hook for
      incremental processing:

          my $options = { hook => \&mysub, }; # I want to do incremental processing
          $bot->linksearch("slashdot.org", 1, undef, $options);
          sub mysub {
              my ($res) = @_;
              foreach my $hashref (@$res) {
                  my $url  = $hashref->{'url'};
                  my $page = $hashref->{'title'};
                  print "$page: $url\n";
              }
          }

    References: Finding external links
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Finding-external-links>,
    API:Exturlusage <https://www.mediawiki.org/wiki/API:Exturlusage>

 purge_page

    Purges the server cache of the specified $page. Returns true on
    success; false on failure. Pass an array reference to purge multiple
    pages.

    If you really care, a true return value is the number of pages
    successfully purged. You could check that it is the same as the number
    you wanted to purge - maybe some pages don't exist, or you passed
    invalid titles, or you aren't allowed to purge the cache:

        my @to_purge = ('Main Page', 'A', 'B', 'C', 'Very unlikely to exist');
        my $size = scalar @to_purge;
    
        print "all-at-once:\n";
        my $success = $bot->purge_page(\@to_purge);
    
        if ($success == $size) {
            print "@to_purge: OK ($success/$size)\n";
        }
        else {
            my $missed = @to_purge - $success;
            print "We couldn't purge $missed pages (list was: "
                . join(', ', @to_purge)
                . ")\n";
        }
    
        # OR
        print "\n\none-at-a-time:\n";
        foreach my $page (@to_purge) {
            my $ok = $bot->purge_page($page);
            print "$page: $ok\n";
        }

    References: Purging the server cache
    <https://github.com/MediaWiki-Bot/MediaWiki-Bot/wiki/Purging-the-server-cache>,
    API:Purge <https://www.mediawiki.org/wiki/API:Purge>

 get_namespace_names

        my %namespace_names = $bot->get_namespace_names();

    Returns a hash linking the namespace id, such as 1, to its named
    equivalent, such as "Talk".

    References: API:Meta#siteinfo
    <https://www.mediawiki.org/wiki/API:Meta#siteinfo_.2F_si>

 image_usage

    Gets a list of pages which include a certain $image. Include the File:
    namespace prefix to avoid incurring an extra round-trip (which will
    also emit a deprecation warnings).

    Additional parameters are:

      * A namespace number to fetch results from (or an arrayref of
      multiple namespace numbers)

      * One of all, redirect, or nonredirects.

      * $options is a hashref as described in the section for "linksearch".

        my @pages = $bot->image_usage("File:Albert Einstein Head.jpg");

    Or, make use of the "Options hashref" to do incremental processing:

        $bot->image_usage("File:Albert Einstein Head.jpg",
            undef, undef,
            { hook=>\&mysub, max=>5 }
        );
        sub mysub {
            my $res = shift;
            foreach my $page (@$res) {
                my $title = $page->{'title'};
                print "$title\n";
            }
        }

    References: API:Imageusage
    <https://www.mediawiki.org/wiki/API:Imageusage>

 global_image_usage($image, $results, $filterlocal)

    Returns an array of hashrefs of data about pages which use the given
    image.

        my @data = $bot->global_image_usage('File:Albert Einstein Head.jpg');

    The keys in each hashref are title, url, and wiki. $results is the
    maximum number of results that will be returned (not the maximum number
    of requests that will be sent, like max in the "Options hashref"); the
    default is to attempt to fetch 500 (set to 0 to get all results).
    $filterlocal will filter out local uses of the image.

    References: Extension:GlobalUsage#API
    <https://www.mediawiki.org/wiki/Extension:GlobalUsage#API>

 links_to_image

    A backward-compatible call to "image_usage". You can provide only the
    image title.

    This method is deprecated, and will emit deprecation warnings.

 is_blocked

        my $blocked = $bot->is_blocked('User:Mike.lifeguard');

    Checks if a user is currently blocked.

    References: API:Blocks <https://www.mediawiki.org/wiki/API:Blocks>

 test_blocked

    Retained for backwards compatibility. Use "is_blocked" for clarity.

    This method is deprecated, and will emit deprecation warnings.

 test_image_exists

    Checks if an image exists at $page.

      * FILE_NONEXISTENT (0) means "Nothing there"

      * FILE_LOCAL (1) means "Yes, an image exists locally"

      * FILE_SHARED (2) means "Yes, an image exists on Commons
      <http://commons.wikimedia.org>"

      * FILE_PAGE_TEXT_ONLY (3) means "No image exists, but there is text
      on the page"

    If you pass in an arrayref of images, you'll get out an arrayref of
    results.

        use MediaWiki::Bot::Constants;
        my $exists = $bot->test_image_exists('File:Albert Einstein Head.jpg');
        if ($exists == FILE_NONEXISTENT) {
            print "Doesn't exist\n";
        }
        elsif ($exists == FILE_LOCAL) {
            print "Exists locally\n";
        }
        elsif ($exists == FILE_SHARED) {
            print "Exists on Commons\n";
        }
        elsif ($exists == FILE_PAGE_TEXT_ONLY) {
            print "Page exists, but no image\n";
        }

    References: API:Properties#imageinfo
    <https://www.mediawiki.org/wiki/API:Properties#imageinfo_.2F_ii>

 get_pages_in_namespace

        $bot->get_pages_in_namespace($namespace, $limit, $options_hashref);

    Returns an array containing the names of all pages in the specified
    namespace. The $namespace_id must be a number, not a namespace name.

    Setting $page_limit is optional, and specifies how many items to
    retrieve at once. Setting this to 'max' is recommended, and this is the
    default if omitted. If $page_limit is over 500, it will be rounded up
    to the next multiple of 500. If $page_limit is set higher than you are
    allowed to use, it will silently be reduced. Consider setting key 'max'
    in the "Options hashref" to retrieve multiple sets of results:

        # Gotta get 'em all!
        my @pages = $bot->get_pages_in_namespace(6, 'max', { max => 0 });

    References: API:Allpages <https://www.mediawiki.org/wiki/API:Allpages>

 count_contributions

        my $count = $bot->count_contributions($user);

    Uses the API to count $user's contributions.

    References: API:Users <https://www.mediawiki.org/wiki/API:Users>

 timed_count_contributions

        ($timed_edits_count, $total_count) = $bot->timed_count_contributions($user, $days);

    Uses the API to count $user's contributions in last number of $days and
    total number of user's contributions (if needed).

    Example: If you want to get user contribs for last 30 and 365 days, and
    total number of edits you would write something like this:

        my ($last30days, $total) = $bot->timed_count_contributions($user, 30);
        my $last365days = $bot->timed_count_contributions($user, 365);

    You could get total number of edits also by separately calling
    count_contributions like this:

        my $total = $bot->count_contributions($user);

    and use timed_count_contributions only in scalar context, but that
    would mean one more call to server (meaning more server load) of which
    you are excused as timed_count_contributions returns array with two
    parameters.

    References: Extension:UserDailyContribs
    <https://www.mediawiki.org/wiki/Extension:UserDailyContribs>

 last_active

        my $latest_timestamp = $bot->last_active($user);

    Returns the last active time of $user in YYYY-MM-DDTHH:MM:SSZ.

    References: API:Usercontribs
    <https://www.mediawiki.org/wiki/API:Usercontribs>

 recent_edit_to_page

         my ($timestamp, $user) = $bot->recent_edit_to_page($title);

    Returns timestamp and username for most recent (top) edit to $page.

    References: API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 get_users

        my @recent_editors = $bot->get_users($title, $limit, $revid, $direction);

    Gets the most recent editors to $page, up to $limit, starting from
    $revision and going in $direction.

    References: API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 was_blocked

        for ("Mike.lifeguard", "Jimbo Wales") {
            print "$_ was blocked\n" if $bot->was_blocked($_);
        }

    Returns whether $user has ever been blocked.

    References: API:Logevents
    <https://www.mediawiki.org/wiki/API:Logevents>

 test_block_hist

    Retained for backwards compatibility. Use "was_blocked" for clarity.

    This method is deprecated, and will emit deprecation warnings.

 expandtemplates

        my $expanded = $bot->expandtemplates($title, $wikitext);

    Expands templates on $page, using $text if provided, otherwise loading
    the page text automatically.

    References: API:Parsing wikitext
    <https://www.mediawiki.org/wiki/API:Parsing_wikitext>

 get_allusers

        my @users = $bot->get_allusers($limit, $user_group, $options_hashref);

    Returns an array of all users. Default $limit is 500. Optionally
    specify a $group (like 'sysop') to list that group only. The last
    optional parameter is an "Options hashref".

    References: API:Allusers <https://www.mediawiki.org/wiki/API:Allusers>

 db_to_domain

    Converts a wiki/database name (enwiki) to the domain name
    (en.wikipedia.org).

        my @wikis = ("enwiki", "kowiki", "bat-smgwiki", "nonexistent");
        foreach my $wiki (@wikis) {
            my $domain = $bot->db_to_domain($wiki);
            next if !defined($domain);
            print "$wiki: $domain\n";
        }

    You can pass an arrayref to do bulk lookup:

        my @wikis = ("enwiki", "kowiki", "bat-smgwiki", "nonexistent");
        my $domains = $bot->db_to_domain(\@wikis);
        foreach my $domain (@$domains) {
            next if !defined($domain);
            print "$domain\n";
        }

    References: Extension:SiteMatrix
    <https://www.mediawiki.org/wiki/Extension:SiteMatrix>

 domain_to_db

        my $db = $bot->domain_to_db($domain_name);

    As you might expect, does the opposite of "domain_to_db": Converts a
    domain name (meta.wikimedia.org) into a database/wiki name (metawiki).

    References: Extension:SiteMatrix
    <https://www.mediawiki.org/wiki/Extension:SiteMatrix>

 diff

    This allows retrieval of a diff from the API. The return is a scalar
    containing the HTML table of the diff. Options are passed as a hashref
    with keys:

      * title is the title to use. Provide either this or revid.

      * revid is any revid to diff from. If you also specified title, only
      title will be honoured.

      * oldid is an identifier to diff to. This can be a revid, or the
      special values 'cur', 'prev' or 'next'

    References: API:Properties#revisions
    <https://www.mediawiki.org/wiki/API:Properties#revisions_.2F_rv>

 prefixindex

    This returns an array of hashrefs containing page titles that start
    with the given $prefix. The hashref has keys 'title' and 'redirect'
    (present if the page is a redirect, not present otherwise).

    Additional parameters are:

      * One of all, redirects, or nonredirects

      * A single namespace number (unlike linksearch etc, which can accept
      an arrayref of numbers).

      * $options_hashref as described in "Options hashref".

        my @prefix_pages = $bot->prefixindex("User:Mike.lifeguard");
        # Or, the more efficient equivalent
        my @prefix_pages = $bot->prefixindex("Mike.lifeguard", 2);
        foreach my $hashref (@pages) {
            my $title = $hashref->{'title'};
            if $hashref->{'redirect'} {
                print "$title is a redirect\n";
            }
            else {
                print "$title\n is not a redirect\n";
            }
        }

    References: API:Allpages <https://www.mediawiki.org/wiki/API:Allpages>

 search

    This is a simple search for your $search_term in page text. It returns
    an array of page titles matching.

    Additional optional parameters are:

      * A namespace number to search in, or an arrayref of numbers (default
      is the main namespace)

      * $options_hashref is a hashref as described in "Options hashref":

        my @pages = $bot->search("Mike.lifeguard", 2);
        print "@pages\n";

    Or, use a callback for incremental processing:

        my @pages = $bot->search("Mike.lifeguard", 2, { hook => \&mysub });
        sub mysub {
            my ($res) = @_;
            foreach my $hashref (@$res) {
                my $page = $hashref->{'title'};
                print "$page\n";
            }
        }

    References: API:Search <https://www.mediawiki.org/wiki/API:Search>

 get_log

    This fetches log entries, and returns results as an array of hashes.
    The first parameter is a hashref with keys:

      * type is the log type (block, delete...)

      * user is the user who performed the action. Do not include the User:
      prefix

      * target is the target of the action. Where an action was performed
      to a page, it is the page title. Where an action was performed to a
      user, it is User:$username.

    The second is the familiar "Options hashref".

        my $log = $bot->get_log({
                type => 'block',
                user => 'User:Mike.lifeguard',
            });
        foreach my $entry (@$log) {
            my $user = $entry->{'title'};
            print "$user\n";
        }
    
        $bot->get_log({
                type => 'block',
                user => 'User:Mike.lifeguard',
            },
            { hook => \&mysub, max => 10 }
        );
        sub mysub {
            my ($res) = @_;
            foreach my $hashref (@$res) {
                my $title = $hashref->{'title'};
                print "$title\n";
            }
        }

    References: API:Logevents
    <https://www.mediawiki.org/wiki/API:Logevents>

 is_g_blocked

        my $is_globally_blocked = $bot->is_g_blocked('127.0.0.1');

    Returns what IP/range block currently in place affects the IP/range.
    The return is a scalar of an IP/range if found (evaluates to true in
    boolean context); undef otherwise (evaluates false in boolean context).
    Pass in a single IP or CIDR range.

    References: Extension:GlobalBlocking
    <https://www.mediawiki.org/wiki/Extension:GlobalBlocking/API>

 was_g_blocked

        print "127.0.0.1 was globally blocked\n" if $bot->was_g_blocked('127.0.0.1');

    Returns whether an IP/range was ever globally blocked. You should
    probably call this method only when your bot is operating on Meta -
    this method will warn if not.

    References: API:Logevents
    <https://www.mediawiki.org/wiki/API:Logevents>

 was_locked

        my $was_locked = $bot->was_locked('Mike.lifeguard');

    Returns whether a user was ever locked. You should probably call this
    method only when your bot is operating on Meta - this method will warn
    if not.

    References: API:Logevents
    <https://www.mediawiki.org/wiki/API:Logevents>

 get_protection

    Returns data on page protection as a array of up to two hashrefs. Each
    hashref has a type, level, and expiry. Levels are 'sysop' and
    'autoconfirmed'; types are 'move' and 'edit'; expiry is a timestamp.
    Additionally, the key 'cascade' will exist if cascading protection is
    used.

        my $page = 'Main Page';
        $bot->edit({
            page    => $page,
            text    => rand(),
            summary => 'test',
        }) unless $bot->get_protection($page);

    You can also pass an arrayref of page titles to do bulk queries:

        my @pages = ('Main Page', 'User:Mike.lifeguard', 'Project:Sandbox');
        my $answer = $bot->get_protection(\@pages);
        foreach my $title (keys %$answer) {
            my $protected = $answer->{$title};
            print "$title is protected\n" if $protected;
            print "$title is unprotected\n" unless $protected;
        }

    References: API:Properties#info
    <https://www.mediawiki.org/wiki/API:Properties#info_.2F_in>

 is_protected

    This is a synonym for "get_protection", which should be used in
    preference.

    This method is deprecated, and will emit deprecation warnings.

 patrol

        $bot->patrol($rcid);

    Marks a page or revision identified by the $rcid as patrolled. To mark
    several RCIDs as patrolled, you may pass an arrayref of them. Returns
    false and sets $bot->{error} if the account cannot patrol.

    References: API:Patrol <https://www.mediawiki.org/wiki/API:Patrol>

 email

        $bot->email($user, $subject, $body);

    This allows you to send emails through the wiki. All 3 of $user
    (without the User: prefix), $subject and $body are required. If $user
    is an arrayref, this will send the same email (subject and body) to all
    users.

    References: API:Email <https://www.mediawiki.org/wiki/API:Email>

 top_edits

    Returns an array of the page titles where the $user is the latest
    editor. The second parameter is the familiar $options_hashref.

        my @pages = $bot->top_edits("Mike.lifeguard", {max => 5});
        foreach my $page (@pages) {
            $bot->rollback($page, "Mike.lifeguard");
        }

    Note that accessing the data with a callback happens before filtering
    the top edits is done. For that reason, you should use "contributions"
    if you need to use a callback. If you use a callback with top_edits(),
    you will not necessarily get top edits returned. It is only safe to use
    a callback if you check that it is a top edit:

        $bot->top_edits("Mike.lifeguard", { hook => \&rv });
        sub rv {
            my $data = shift;
            foreach my $page (@$data) {
                if (exists($page->{'top'})) {
                    $bot->rollback($page->{'title'}, "Mike.lifeguard");
                }
            }
        }

    References: API:Usercontribs
    <https://www.mediawiki.org/wiki/API:Usercontribs>

 contributions

        my @contribs = $bot->contributions($user, $namespace, $options, $from, $to);

    Returns an array of hashrefs of data for the user's contributions.
    $namespace can be an arrayref of namespace numbers. $options can be
    specified as in "linksearch". $from and $to are optional timestamps.
    ISO 8601 date and time is recommended: 2001-01-15T14:56:00Z, see
    https://www.mediawiki.org/wiki/Timestamp for all possible formats. Note
    that $from (=ucend) has to be before $to (=ucstart), unlike direct API
    access.

    Specify an arrayref of users to get results for multiple users.

    References: API:Usercontribs
    <https://www.mediawiki.org/wiki/API:Usercontribs>

 upload

        $bot->upload({ data => $file_contents, summary => 'uploading file' });
        $bot->upload({ file => $file_name,     title   => 'Target filename.png' });

    Upload a file to the wiki. Specify the file by either giving the
    filename, which will be read in, or by giving the data directly.

    References: API:Upload <https://www.mediawiki.org/wiki/API:Upload>

 upload_from_url

    Upload file directly from URL to the wiki. Specify URL, the new
    filename and summary. Summary and new filename are optional.

        $bot->upload_from_url({
            url => 'http://some.domain.ext/pic.png',
            title => 'Target_filename.png',
            summary => 'uploading new pic',
        });

    If on your target wiki is enabled uploading from URL, meaning
    $wgAllowCopyUploads is set to true in LocalSettings.php and you have
    appropriate user rights, you can use this function to upload files to
    your wiki directly from remote server.

    References: API:Upload#Uploading_from_URL
    <https://www.mediawiki.org/wiki/API:Upload#Uploading_from_URL>

 usergroups

    Returns a list of the usergroups a user is in:

        my @usergroups = $bot->usergroups('Mike.lifeguard');

    References: API:Users <https://www.mediawiki.org/wiki/API:Users>

 get_mw_version

    Returns a hash ref with the MediaWiki version. The hash ref contains
    the keys major, minor, patch, and string. Returns undef on errors.

        my $mw_version = $bot->get_mw_version;
    
        # get version as string
        my $mw_ver_as_string = $mw_version->{'major'} . '.' . $mw_version->{'minor'};
        if(defined $mw_version->{'patch'}){
            $mw_ver_as_string .= '.' . $mw_version->{'patch'};
        }
    
        # or simply
        my $mw_ver_as_string = $mw_version->{'string'};

    References: API:Siteinfo <https://www.mediawiki.org/wiki/API:Siteinfo>

 Options hashref

    This is passed through to the lower-level interface MediaWiki::API, and
    is fully documented there.

    The hashref can have 3 keys:

    max

      Specifies the maximum number of queries to retrieve data from the
      wiki. This is independent of the size of each query (how many items
      each query returns). Set to 0 to retrieve all the results.

    hook

      Specifies a coderef to a hook function that can be used to process
      large lists as they come in. When this is used, your subroutine will
      get the raw data. This is noted in cases where it is known to be
      significant. For example, when using a hook with top_edits(), you
      need to check whether the edit is the top edit yourself - your
      subroutine gets results as they come in, and before they're filtered.

    skip_encoding

      MediaWiki's API uses UTF-8 and any 8 bit character string parameters
      are encoded automatically by the API call. If your parameters are
      already in UTF-8 this will be detected and the encoding will be
      skipped. If your parameters for some reason contain UTF-8 data but no
      UTF-8 flag is set (i.e. you did not use the use utf8; pragma) you
      should prevent re-encoding by passing an option skip_encoding => 1.
      For example:

          $category ="Cat\x{e9}gorie:moyen_fran\x{e7}ais"; # latin1 string
          $bot->get_all_pages_in_category($category); # OK
      
          $category = "Cat". pack("U", 0xe9)."gorie:moyen_fran".pack("U",0xe7)."ais"; # unicode string
          $bot->get_all_pages_in_category($category); # OK
      
          $category ="Cat\x{c3}\x{a9}gorie:moyen_fran\x{c3}\x{a7}ais"; # unicode data without utf-8 flag
          # $bot->get_all_pages_in_category($category); # NOT OK
          $bot->get_all_pages_in_category($category, { skip_encoding => 1 }); # OK

      If you need this, it probably means you're doing something wrong.
      Feel free to ask for help.

ERROR HANDLING

    All functions will return undef in any handled error situation. Further
    error data is stored in $bot->{error}->{code} and
    $bot->{error}->{details}.

    Error codes are provided as constants in MediaWiki::Bot::Constants, and
    can also be imported through this module:

        use MediaWiki::Bot qw(:constants);