US11212954B2 - Apparatus and methods for field operations based on historical field operation data - Google Patents

Apparatus and methods for field operations based on historical field operation data Download PDF

Info

Publication number
US11212954B2
US11212954B2 US16/406,830 US201916406830A US11212954B2 US 11212954 B2 US11212954 B2 US 11212954B2 US 201916406830 A US201916406830 A US 201916406830A US 11212954 B2 US11212954 B2 US 11212954B2
Authority
US
United States
Prior art keywords
implement
field
vehicle
guidance line
guidance
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/406,830
Other versions
US20200352082A1 (en
Inventor
Curtis A. Maeder
Qiang R. Liu
Terence D. Pickett
Jeffery J. Adams
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Deere and Co
Original Assignee
Deere and Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Deere and Co filed Critical Deere and Co
Priority to US16/406,830 priority Critical patent/US11212954B2/en
Assigned to DEERE & COMPANY reassignment DEERE & COMPANY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ADAMS, JEFFERY J., LIU, QIANG R., MAEDER, CURTIS A., PICKETT, TERENCE D.
Priority to EP20173207.0A priority patent/EP3735811B1/en
Publication of US20200352082A1 publication Critical patent/US20200352082A1/en
Priority to US17/645,264 priority patent/US11825761B2/en
Application granted granted Critical
Publication of US11212954B2 publication Critical patent/US11212954B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01BSOIL WORKING IN AGRICULTURE OR FORESTRY; PARTS, DETAILS, OR ACCESSORIES OF AGRICULTURAL MACHINES OR IMPLEMENTS, IN GENERAL
    • A01B69/00Steering of agricultural machines or implements; Guiding agricultural machines or implements on a desired track
    • A01B69/007Steering or guiding of agricultural vehicles, e.g. steering of the tractor to keep the plough in the furrow
    • A01B69/008Steering or guiding of agricultural vehicles, e.g. steering of the tractor to keep the plough in the furrow automatic
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01BSOIL WORKING IN AGRICULTURE OR FORESTRY; PARTS, DETAILS, OR ACCESSORIES OF AGRICULTURAL MACHINES OR IMPLEMENTS, IN GENERAL
    • A01B69/00Steering of agricultural machines or implements; Guiding agricultural machines or implements on a desired track
    • A01B69/001Steering by means of optical assistance, e.g. television cameras
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01BSOIL WORKING IN AGRICULTURE OR FORESTRY; PARTS, DETAILS, OR ACCESSORIES OF AGRICULTURAL MACHINES OR IMPLEMENTS, IN GENERAL
    • A01B69/00Steering of agricultural machines or implements; Guiding agricultural machines or implements on a desired track
    • A01B69/007Steering or guiding of agricultural vehicles, e.g. steering of the tractor to keep the plough in the furrow
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01BSOIL WORKING IN AGRICULTURE OR FORESTRY; PARTS, DETAILS, OR ACCESSORIES OF AGRICULTURAL MACHINES OR IMPLEMENTS, IN GENERAL
    • A01B79/00Methods for working soil
    • AHUMAN NECESSITIES
    • A01AGRICULTURE; FORESTRY; ANIMAL HUSBANDRY; HUNTING; TRAPPING; FISHING
    • A01BSOIL WORKING IN AGRICULTURE OR FORESTRY; PARTS, DETAILS, OR ACCESSORIES OF AGRICULTURAL MACHINES OR IMPLEMENTS, IN GENERAL
    • A01B79/00Methods for working soil
    • A01B79/005Precision agriculture
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0212Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0268Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means
    • G05D1/0274Control of position or course in two dimensions specially adapted to land vehicles using internal positioning means using mapping information stored in a memory device

Definitions

  • This disclosure relates generally to agricultural techniques, and, more particularly, to apparatus and methods for field operations based on historical field operation data.
  • Agricultural vehicles may perform semi-autonomous or fully-autonomous operations on fields.
  • Agricultural vehicles perform operations using implements including planting implements, spraying implements, harvesting implements, fertilizing implements, strip/till implements, etc.
  • An example apparatus disclosed herein includes a guidance line generator to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field.
  • the example apparatus further includes a drive commander to cause the vehicle to traverse the field along the guidance line and an implement commander to cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
  • An example non-transitory computer readable storage medium disclosed herein includes instructions that, when executed, cause a processor to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field.
  • the instructions when executed, further cause the processor to cause the vehicle to traverse the field along the guidance line, and cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
  • An example method disclosed herein includes generating a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field.
  • the example method further includes causing the vehicle to traverse the field along the guidance line and causing the implement to perform the second operation as the vehicle traverses the field along the guidance line.
  • FIG. 1 is a schematic of an example vehicle constructed in accordance with the teachings of this disclosure.
  • FIG. 2 is a schematic of an example controller of the vehicle of FIG. 1 .
  • FIG. 3 is a schematic illustrating an example coverage-based mapping technique and an example row-based mapping technique.
  • FIG. 4A is an example field map generated using an example four-row implement.
  • FIG. 4B is an example first field map traversal schematic using an example twelve-row sprayer in a subsequent operation on the field corresponding to the field map of FIG. 4A .
  • FIG. 4C is an example second field map traversal schematic using an example eight-row sprayer in a subsequent operation on the field corresponding to the field map of FIG. 4A .
  • FIG. 4D is an example third field map traversal schematic using an example four-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A .
  • FIG. 4E is an example fourth field map traversal schematic using an example two-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A .
  • FIG. 4F is an example fifth field map traversal schematic using an example one-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A .
  • FIG. 5A is an enlarged view of the example field map of FIG. 4A generated using a four-row implement.
  • FIG. 5B is an example sixth field map traversal schematic using an example single row harvester with an example haulage vehicle on the field corresponding to the field map of FIG. 5A .
  • FIG. 5C is an example seventh field map traversal schematic using the example single row harvester of FIG. 5B on the field corresponding to the field map of FIG. 5A .
  • FIG. 5D is an example eighth field map traversal schematic using the example haulage vehicle of FIG. 5B on the field corresponding to the field map of FIG. 5A .
  • FIG. 6A is an example ninth field map traversal schematic illustrating an original operation and a subsequent operation executed on a field map with a narrow guess row.
  • FIG. 6B is an example tenth field map traversal schematic illustrating another original operation and another subsequent operation executed on a field map with an accurate guess row.
  • FIG. 7 is an example run screen to be displayed on an example user interface of the vehicle of FIG. 1 .
  • FIG. 8 is a flowchart representative of example machine readable instructions that may be executed to implement the controller of FIGS. 1 and 2 to perform an operation on a field using a field map.
  • FIG. 9 is a flowchart representative of example machine readable instructions that may be executed to implement the controller of FIGS. 1 and 2 to generate a first operation map.
  • FIG. 10 is a flowchart representative of example machine readable instructions that may be executed to implement the controllers of FIGS. 1 and 2 to automatically guide a vehicle along guidance line(s) and perform an operation on a field.
  • FIG. 11 is a block diagram of an example processing platform structured to execute the example machine readable instructions of FIGS. 8-10 to implement the controller of FIGS. 1 and 2 .
  • the figures are not to scale. In general, the same reference numbers will be used throughout the drawing(s) and accompanying written description to refer to the same or like parts.
  • Descriptors “first,” “second,” “third,” etc. are used herein when identifying multiple elements or components which may be referred to separately. Unless otherwise specified or understood based on their context of use, such descriptors are not intended to impute any meaning of priority, physical order or arrangement in a list, or ordering in time but are merely used as labels for referring to multiple elements or components separately for ease of understanding the disclosed examples.
  • the descriptor “first” may be used to refer to an element in the detailed description, while the same element may be referred to in a claim with a different descriptor such as “second” or “third.” In such instances, it should be understood that such descriptors are used merely for ease of referencing multiple elements or components.
  • Agricultural vehicles perform numerous operations on fields using different implements.
  • a planter implement may plant crops at the beginning of a season.
  • a sprayer implement may subsequently spray the crops, and/or a fertilizer implement may subsequently fertilize the crops.
  • Different vehicles may be used for individual operations, or different implements may be attached to a same vehicle (e.g., a tractor) to perform different operations.
  • a planting operation, a strip/till operation, or a fertilizer operation may be performed first, followed by a spraying operation, a harvesting operation (e.g., a combine or forage harvester), a side/top dress fertilizing operation, or any other field operation.
  • an “operation” can include planting, spraying, fertilizing, and/or performing any other task on a field.
  • operations performed on the field are additionally or alternatively referred to as “treatments.” Further, performing such operations on a field can be referred to as “operating upon” the field, “treating” the field, or “working” the field.
  • an operation may include installation of permanent assets, such as beds or drip irrigation lines. While example orderings of operations (e.g., planting followed by fertilizing) may be discussed herein, operations may be performed in any order.
  • Automated vehicles move by following guidance lines.
  • Conventional guidance lines are generated by defining a first point (e.g., an “A” point) and a second point (e.g., a “B” point) and generating a line or curve connecting the first point and the second point (e.g., an A-B line, or an A-B curve).
  • Automatically generating point-to-point guidance lines across a field dimension fails to account for topography of the field and specific locations of crop rows.
  • Most operations are subsequent operations performed on a field (e.g., most operations are performed on existing crops that are planted in the field).
  • subsequent operations may be manually performed by an operator or automatically performed while accepting damage to the crop rows.
  • manual operations the operator analyzes the crop rows and attempts to position the agricultural vehicle to operate on the crop rows to minimize damage to crops or waste of product (e.g., spray, fertilizer, etc.).
  • subsequent field operations are performed automatically based on planned or expected crop row locations.
  • sources of error such as estimation of implement drift in curved paths, drift due to slopes in the terrain, uneven draft loads across the width of an implement which cause the implement to pull off-center, disturbances due to soil surface roughness, and/or other sources of error, planned crop row locations are often significantly different than actual crop row locations.
  • agricultural vehicle systems attempt to sense locations of crop rows as they operate.
  • real-time sensing is insufficient to avoid crop damage, and fails to reliably address unique crop row locations.
  • conventional agricultural vehicle systems may utilize an overlap, where crop rows at the edge of the implement get operated upon twice in two successive passes, resulting in inefficiency and waste.
  • the planter When a planting operation is performed, after completing planting of one set of rows (e.g., one-pass of the field in a direction), the planter must return in a new pass in the opposite direction. To begin a new set of rows, the operator must estimate a starting position for the first row at an end of the planting implement. Specifically, the operator must try and line up the vehicle such that the first new row is spaced apart from the nearest existing crop row at the same spacing as exists between the rest of the rows. Such edge rows are referred to as “guess rows,” since the operator must guess the appropriate spacing to plant the crop row. This is a very challenging task, which often results in the first row having too narrow a gap relative to the nearest existing row, or too large a gap.
  • an operator may have difficulty performing an operation on the guess rows. Additionally, the size of the guess row can vary from pass to pass (e.g., between traversals) over the field, introducing additional variability that makes subsequent operations increasingly difficult. In some conventional implementations, an operator will merely drive over and sacrifice these “guess rows” during subsequent operations.
  • Example methods, apparatus, systems, and articles of manufacture (e.g., physical storage media) disclosed herein generate guidance lines for automatic execution of subsequent operations on a field based on field maps determined during previous operations. Some examples disclosed herein generate field maps based on location data (e.g., GPS and inertial data) and sensor data from implements to accurately represent crop rows in field maps.
  • Example methods, apparatus, systems, and articles of manufacture (e.g., physical storage media) disclosed herein enable accurate navigation along guidance lines to accurately operate upon a field regardless of the terrain of the field and/or the size of implement utilized. Examples disclosed herein accurately operate upon guess rows by generating guidance lines specific to locations of the guess rows, eliminating potential operator error.
  • Performing subsequent operations based on field maps improves alignment relative to established row locations and minimizes damage to crops, as the wheels of the agricultural vehicle can be positioned to avoid running over rows.
  • fields can be operated upon (e.g., harvested, sprayed, fertilized, etc.) faster since the guidance system can anticipate upcoming changes in the terrain and/or changes in crop row locations, rather than relying upon an operator to make adjustments and/or a sensor to sense changes in the terrain.
  • some crops, such as corn require highly precise positioning of the implement (e.g., a harvesting head, a platform for small grain harvesting, etc.) during harvesting. Utilization of accurate mapping data to generate precise guidance lines helps ensure that crops are operated upon safely and efficiently.
  • Example techniques disclosed herein enable an operator to begin a subsequent operation at any location in a field, as mapping data can be utilized to automatically generate guidance lines for any location. Further, example techniques disclosed herein enable a coordinated work system in which a plurality of agricultural vehicles complete an operation by each generating guidance lines based off shared mapping data and knowledge of other vehicles' planned operations. Similarly, a plurality of agricultural vehicles can generate the field map during an initial operation (e.g., a planting operation).
  • an initial operation e.g., a planting operation
  • Example techniques disclosed herein further utilize individual section and/or nozzle control to ensure that if a portion of the implement is passing over a crop row which has already been operated upon (e.g., already sprayed), this portion of the implement does not perform a redundant operation.
  • FIG. 1 is a schematic of an example vehicle 102 constructed in accordance with the teachings of this disclosure.
  • the example vehicle 102 includes an example main body 104 including an example cab 106 from which an operator can operate the vehicle 102 .
  • the vehicle 102 may not include the cab 106 .
  • the vehicle 102 further includes an example implement 108 .
  • the implement 108 is a planting implement.
  • the implement 108 is a twelve-row planter.
  • the implement 108 plants crops in example crop rows 110 a - l .
  • the implement 108 may be separable from the vehicle 102 , and exchanged with another implement of a same or different implement type.
  • the implement may be a sprayer implement, a fertilizer implement, a tilling implement, a harvesting implement, and/or any other type of implement to perform an operation on a field.
  • the implement 108 is not separable from the vehicle 102 , and multiple vehicles may be used to perform different operations on a field. While the implement 108 of FIG.
  • the implement 108 may have any operational width (e.g., four rows, eight rows, sixteen rows, etc.).
  • the operational width of the implement 108 defines the number of rows that can be operated upon (e.g., planted, tilled, sprayed, etc.) in a single pass of the vehicle 102 over a field. While the implement 108 of the illustrated example is capable of planting twelve rows simultaneously, in some examples, only some of the rows are planted. For example, if a right-side of the implement 108 is traveling over a portion of a field that has already been planted, the corresponding rows of the implement 108 may be disabled to prevent performing a repeat operation.
  • Individual row portions of the implement 108 of the illustrated example of FIG. 1 include example implement sensors 112 a - l to detect whether the row portions of the implement 108 are performing operations. While only the implement sensors 112 a - e are visible due to the angle of the perspective view of FIG. 1 , additional implement sensors 112 f - l are understood to be mounted to portions of the implement 108 in alignment with crop rows 110 f - l .
  • the implement sensors 112 a - 1 are each positioned on one of the portions of the implement 108 to detect whether operations are performed on individual rows.
  • the implement sensors 112 a - 1 are switches which activate as a product passes through the implement portion (e.g., as seed exits the implement 108 ). Any type of sensor may be utilized to determine whether a portion of the implement 108 corresponding to a specific row is performing an operation.
  • the implement sensors 112 a - 1 communicate with an example controller 116 .
  • the implement 108 of the illustrated example of FIG. 1 includes an example location sensor 114 .
  • the location sensor 114 of the illustrated example is a global positioning system (GPS) sensor combined with an inertial sensor.
  • GPS global positioning system
  • the inertial sensor enables higher precision location data by enabling compensation for tilt of the vehicle 102 . For example, if the vehicle 102 is tilted on a hill, the inertial sensor can compensate for this tilt to enable accurate determination of a location of the vehicle 102 , and more specifically, accurate determination of locations of portions of the implement 108 .
  • the example location sensor 114 of the illustrated example is mounted to the implement 108 to provide more accurate ground-level location data for subsequent determination of crop row locations.
  • the location sensor 114 is mounted to another part of the vehicle 102 (e.g., the main body 104 , the cab 106 , etc.). In some examples, the location sensor 114 includes multiple location sensors. In some such examples, the vehicle 102 includes the location sensor 114 as a first sensor on the main body 104 to provide a location for the vehicle 102 generally, and a second sensor on the implement 108 to provide ground-level location data. The location sensor 114 communicates to an example controller 116 .
  • the example controller 116 of the illustrated example of FIG. 1 analyzes data from one or more sensors and enables automated operation of the vehicle 102 .
  • the controller 116 of the illustrated example generates field maps and utilizes such field maps to automatically generate guidance lines based on the field maps and the type of implement in use by the vehicle 102 .
  • the controller 116 accesses implement sensor data from the implement sensors 112 a - l and location data from the location sensor 114 to generate field maps.
  • the controller 116 can determine locations of the crop rows 110 a - l based on the implement sensor data and the location data.
  • the controller 116 of the illustrated example issues commands to cause movement of the vehicle 102 by commanding an example drive controller 118 .
  • controller 116 of the illustrated example issues commands to cause operation of the implement 108 by commanding an example implement controller 120 to perform an operation based on a field map and a parameter of the implement 108 .
  • the controller 116 of the illustrated example further controls an example user interface 122 , which can provide mapping data to an operator and solicit input from the operator. Further detail pertaining to the controller 116 is illustrated and described in association with FIG. 2 .
  • the example drive controller 118 of the illustrated example of FIG. 1 causes movement of the vehicle 102 .
  • the controller 116 can issue commands to the drive controller 118 for the vehicle 102 to move according to a guidance line generated by the controller 116 .
  • the drive controller 118 responds to these commands and causes movement of the vehicle 102 (e.g., movement of the wheels, steering adjustments, etc.).
  • the example implement controller 120 of the illustrated example of FIG. 1 causes the implement 108 to perform an operation on the field.
  • the implement controller 120 can cause the implement 108 to plant, spray, harvest, and/or perform any other operation.
  • the implement controller 120 causes portions of the implement 108 to perform an operation, based on data from the controller 116 indicating that some of the rows under the implement 108 have already been operated upon (e.g., as determined based on mapping data).
  • the drive controller 118 and/or the implement controller 120 may be integrated into the controller 116 .
  • FIG. 2 is a schematic of the controller 116 of the vehicle 102 of FIG. 1 .
  • the controller 116 includes an example field map handler 202 , an example location data analyzer 204 , an example field map generator 206 , an example map data store 208 , an example field map analyzer 210 , an example user command accessor 212 , an example user interface configurator 214 , an example equipment parameter determiner 216 , an example guidance line generator 218 , an example implement commander 220 , an example drive commander 222 , and an example operation performance tracker 224 .
  • the example field map handler 202 of the illustrated example of FIG. 2 analyzes location data and generates and stores field maps based on the location data.
  • the field map handler 202 receives GPS data and/or inertial data from the location sensor 114 . If a field map already exists for a location indicated in the location data (e.g., indicating the current operation is a subsequent operation), then the field map handler 202 can pass the location data (e.g., the GPS data and/or the inertial data) to the field map analyzer 210 , which can retrieve the field map from the map data store 208 . Conversely, if a map has not been generated, the field map generator 206 can generate a new field map based on the location data.
  • the field map handler 202 of the illustrated example includes an example location data analyzer 204 , an example field map generator 206 , and an example map data store 208 .
  • the example location data analyzer 204 of the illustrated example of FIG. 2 analyzes location data from the location sensor 114 .
  • the location data analyzer 204 receives GPS data and inertial data from the location sensor 114 , and analyzes the GPS data and inertial data to determine precise locations of the vehicle 102 and/or the implement 108 .
  • the location data from the location sensor 114 already directly indicates locations of the vehicle 102 and/or the implement 108 .
  • the location data analyzer 204 of the illustrated example determines, based on the location data, locations of portions of the implement 108 corresponding to individual ones of the crop rows 110 .
  • the location data analyzer 204 queries the map data store 208 to determine whether a field map exists for the current location of the vehicle 102 . In some examples, the location data analyzer 204 determines whether a field map exists within a specified time period (e.g., a field map from the past six months, a field map from the past year, etc.), as outdated maps (outside the time period) may no longer be accurate if a field has been re-planted. In some examples, the location data analyzer 204 communicates location data to the field map analyzer 210 if a field map corresponding to the location data already exists.
  • a specified time period e.g., a field map from the past six months, a field map from the past year, etc.
  • the example field map generator 206 of the illustrated example of FIG. 2 generates field maps based on location data from the location sensor 114 .
  • the location data analyzer 204 analyzes the location data to determine locations of the vehicle 102 and/or portions of the implement 108 , which are then received by the field map generator 206 .
  • the field map generator 206 of the illustrated example creates a new map when location data is received for a location for which no map data, or no recent map data (e.g., from the current agricultural season) exists.
  • the field map generator 206 of the illustrated example stores (a) location data including the guidance line which the vehicle 102 followed while performing the initial operation (e.g., planting, tilling, etc.). (b) locations of portions of the implement 108 during the operation, and/or (c) as data from the implement sensors 112 a - l indicating whether an operation was performed by portions of the implement 108 .
  • the operation performance tracker 224 updates the field map generated by the field map generator 206 based on data from the implement sensors 112 a - l .
  • the operation performance tracker 224 may indicate that three portions of the implement 108 were not utilized during a pass through a field.
  • the field map generator 206 updates the field map to indicate the operation did not occur on the crop rows corresponding to these three portions of the implement 108 during the specified pass through the field.
  • the field map generated by the field map generator 206 can indicate (a) precise locations (e.g., accounting for displacements in all three-axes) of the vehicle's guidance line that was followed during an initial operation, (b) precise locations of portions of the implement 108 corresponding to crop rows, and/or (c) indications of whether portions of the implement 108 were utilized during the initial operation.
  • the field map generator 206 stores the field map in the map data store 208 and/or another location accessible to the controller 116 .
  • field maps may be stored at a location accessible via a network, thereby enabling multiple vehicles to access field maps (e.g., in case multiple vehicles perform a coordinated operation, in case different vehicles are utilized in subsequent operations, etc.).
  • the field map generator 206 may combine mapping data (e.g., location data, implement operation data, etc.) from multiple vehicles into a common map for a field.
  • the field map generator 206 may be implemented on a network and/or at a central computing system.
  • the field map generator 206 of one vehicle, such as the vehicle 102 serves as a primary field map generator when multiple vehicles are performing a coordinated operation.
  • individual vehicles performing a coordinated operation may calculate guidance lines and perform other guidance calculations independently, but based upon the same set of rules, such that the vehicles operate predictably and will be coordinated in performing the operation.
  • the map data store 208 of the illustrated example of FIG. 2 stores field maps, which can be subsequently accessed to enable automated operation of the vehicle 102 (e.g., via generation and utilization of accurate, crop row-specific guidance lines).
  • the map data store 208 is accessible via a network.
  • the map data store 208 may be implemented by a volatile memory (e.g., a Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM), etc.) and/or a nonvolatile memory (e.g., flash memory).
  • SDRAM Synchronous Dynamic Random Access Memory
  • DRAM Dynamic Random Access Memory
  • RDRAM RAMBUS Dynamic Random Access Memory
  • nonvolatile memory e.g., flash memory
  • the map data store 208 may additionally or alternatively be implemented by one or more double data rate (DDR) memories, such as DDR, DDR2, DDR3, mobile DDR (mDDR), etc.
  • the map data store 208 may additionally or alternatively be implemented by one or more mass storage devices such as hard disk drive(s), compact disk drive(s) digital versatile disk drive(s), etc. While in the illustrated example the map data store 208 is illustrated as a single database, the map data store 208 may be implemented by any number and/or type(s) of databases.
  • the data stored in the map data store 208 may be in any data format such as, for example, binary data, comma delimited data, tab delimited data, structured query language (SQL) structures, etc.
  • the location data analyzer 204 , the field map generator 206 , the field map analyzer 210 , and/or the operation performance tracker 224 may access and/or modify field maps stored in the map data store 208 .
  • the example field map analyzer 210 of the illustrated example of FIG. 2 analyzes field maps stored in the map data store 208 .
  • the field map analyzer 210 can retrieve field maps based on location data from the location data analyzer 204 .
  • the field map analyzer 210 can retrieve field maps for any location in a field, thereby enabling the guidance line generator 218 to generate guidance lines at any location. Therefore, the field map analyzer 210 enables the vehicle 102 to begin an operation at any location in a field and accurately operate on the field based on precise locations of the crop rows 110 a - 1 represented in the field map.
  • the field map analyzer 210 of the illustrated example determines crop row locations based on field maps retrieved from the map data store 208 . In some examples where multiple vehicles perform a coordinated operation, the field map analyzer 210 determines appropriate portions of a field map to communicate with individual vehicles executing operations on portions of the field.
  • the example user command accessor 212 of the illustrated example of FIG. 2 accesses commands issued by an operator.
  • the user command accessor 212 can receive commands entered by an operator via the user interface 122 .
  • the user command accessor 212 can receive commands to perform various tasks, such as initiating or terminating an automated operation on a field, manually adjusting guidance lines for the vehicle 102 to travel along, initiating or terminating collection of data (e.g., location data, operation performance data, etc.), manually adjusting a setting of the implement (e.g., a spray power, a harvest speed, etc.), and/or any other tasks associated with operation of the vehicle 102 .
  • the user command accessor 212 analyzes command(s) entered by an operator and causes implementation of the command(s).
  • the user command accessor 212 can communicate with the operation performance tracker 224 to cease analysis and storage of implement operation data and the location data analyzer 204 to cease analysis and storage of location data (e.g., GPS data, inertial data, etc.).
  • location data e.g., GPS data, inertial data, etc.
  • Example commands which may be received at the user command accessor 212 include those that may be issued from the run screen 702 illustrated and described in association with FIG. 7 .
  • the example user interface configurator 214 of the illustrated example of FIG. 2 configures the user interface 122 .
  • the user interface configurator 214 can communicate data and/or images to be displayed on the user interface 122 .
  • the user interface configurator 214 communicates map data from the field map analyzer 210 and/or the map data store 208 to the user interface configurator 214 to cause the user interface 122 to display a field map, crop row locations, and/or other data included in a field map.
  • the user interface configurator 214 accesses guidance lines generated by the guidance line generator 218 and causes the user interface 122 to display the guidance lines over the field map for a specific operation (e.g., planting, cultivating, spraying, etc.).
  • the user interface configurator 214 communicates operation performance data from the operation performance tracker 224 to be displayed on the user interface 122 .
  • the user interface 122 may display color and/or shading indicating areas in which an operation has been performed.
  • An example user interface configured by the user interface configurator 214 is illustrated and described in association with FIG. 7 .
  • the example equipment parameter determiner 216 of the illustrated example of FIG. 2 determines parameters of the implement 108 .
  • the equipment parameter determiner 216 may determine a type of the implement 108 (e.g., harvester, sprayer, planter, etc.), an operational width of the implement 108 (e.g., one-row, four-rows, eight-rows, etc.), and/or any other parameters for the implement 108 .
  • the equipment parameter determiner 216 automatically receives data from the implement 108 when the implement 108 is connected to the vehicle 102 .
  • the user command accessor 212 accesses commands including parameters association with the implement 108 . For example, an operator may input a type and/or width of the implement into the user interface 122 , which can then be accessed by the user command accessor 212 and communicated to the equipment parameter determiner 216 .
  • the equipment parameter determiner 216 communicates with the guidance line generator 218 to enable generation of guidance lines that are specific to parameters of the implement 108 . For example, if a field to be operated upon includes thirty-six rows according to the field map as analyzed by the field map analyzer 210 , and the equipment parameter determiner 216 determines that the implement 108 can operate on four rows at a time, the guidance line generator 218 may cause nine guidance lines to be generated to ensure all rows are operated upon.
  • the example guidance line generator 218 of the illustrated example of FIG. 2 generates guidance lines that are specific to the implement 108 attached to the vehicle 102 and/or specific to the field map corresponding to a current field being operated upon.
  • the guidance line generator 218 of the illustrated example generates guidance lines to ensure the vehicle 102 moves along a path that allows the implement 108 to perform its designated operation (e.g., planting, spraying, fertilizing, etc.) most efficiently.
  • the guidance line generator 218 generates as few guidance lines as possible to cover an entire field.
  • the guidance line generator 218 generates guidance lines to avoid wheels of the vehicle 102 or wheels of the implement 108 running over crops rows.
  • the equipment parameter determiner 216 receives information pertaining to crop rows identified in field maps by the field map analyzer 210 and information pertaining to equipment parameters from the equipment parameter determiner 216 .
  • the guidance line generator 218 outputs guidance lines to the drive commander 222 and the implement commander 220 to enable the drive commander 222 to follow the guidance lines when the vehicle 102 is operating in an autonomous mode, and to enable the implement commander 220 to foresee the route of the vehicle 102 and cause the implement 108 to perform efficient operations (e.g., not repeating operations on rows that are passed over twice).
  • the guidance line generator 218 of the illustrated example communicates the guidance lines to the user interface configurator 214 to enable an operator to view the guidance lines on the user interface 122 .
  • the implement commander 220 of the illustrated example of FIG. 2 causes the implement 108 to perform operations on the field.
  • the implement commander 220 actuates portions of the implement 108 to cause an operation (e.g., planting, tilling, spraying) to be executed by the portions of the implement 108 .
  • the implement commander 220 utilizes field map data from the field map analyzer 210 and guidance line data from the guidance line generator 218 to cause portions of the implement 108 to perform operations.
  • the implement commander 220 can reference corresponding field map data from the field map analyzer to determine whether the current operation (e.g., corresponding to the current type of implement) has already been performed on rows that will be passed by the implement 108 .
  • the implement 108 can selectively activate portions of the implement 108 to intelligently perform operations on only those rows which have not yet been operated upon.
  • the implement commander 220 of the illustrated example communicates implement control commands to the implement controller 120 .
  • the implement commander 220 can send implement control commands to the implement controller 120 based on manual operator inputs (e.g., implement controls manually actuated by the operator).
  • the drive commander 222 of the illustrated example of FIG. 2 issues drive control commands to the drive controller 118 to cause the vehicle 102 to move. In some examples, the drive commander 222 issues the drive control commands to cause the vehicle 102 to move along a guidance line generated by the guidance line generator 218 . In some examples, the drive commander 222 issues drive control commands based on user commands input into the user interface 122 (e.g., as accessed by the user command accessor 212 ) or commands otherwise issued form another control in the cab 106 . The drive commander 222 of the illustrated example causes components of the vehicle 102 to adjust to terrain represented in the field map analyzed by the field map analyzer 210 while traversing a guidance line.
  • the drive commander 222 may operate autonomously (e.g., cause the vehicle 102 to automatically follow guidance lines generated by the guidance line generator 218 ), while the implement commander 220 may be manually operated (e.g., may be responsive to manual operator inputs), or vice-versa.
  • both the drive commander 222 and the implement commander 220 operate autonomously based on location data, field map data, equipment parameters, guidance lines, and/or implement operation data.
  • the operation performance tracker 224 of the illustrated example of FIG. 2 analyzes implement operation data from the implement sensors 112 a - l .
  • the operation performance tracker 224 of the illustrated example integrates operation performance data into one or more field maps corresponding to a field currently under operation. For example, if the implement operation data indicates that a portion of the crop rows 110 a - c was operated upon, the data from the corresponding implement sensors 112 a - c can be analyzed by the operation performance tracker 224 , and the field map corresponding to a current location (e.g., as determined by the location data analyzer) can be updated to include data indicating these rows have upon operated upon.
  • a current location e.g., as determined by the location data analyzer
  • the implement commander 220 can intelligently operate the implement 108 to avoid performing a repeat operation on these rows 110 a - c .
  • the operation performance tracker 224 may directly modify field maps from the map data store 208 and/or another storage location, or may communicate implement operation data to the field map analyzer 210 , the field map generator 206 , and/or the implement commander 220 .
  • the field map handler 202 enables generation of field maps by analyzing location data using the location data analyzer 204 , generating maps based on the location data using the field map generator 206 , and storing the field maps for subsequent use in the map data store 208 .
  • the field map analyzer 210 can analyze the existing map and determine crop row locations and other data to enable automated navigation and operation throughout the field.
  • the user command accessor 212 can process a variety of commands from an operator via the user interface 122 or another component in the cab 106 .
  • the user interface configurator 214 communicates user interface data to the user interface 122 such that field maps and other data can be displayed on the user interface 122 for the operator.
  • the equipment parameter determiner 216 determines parameters (e.g., a type of implement, an operational width of an implement, etc.) of the implement 108 .
  • the guidance line generator 218 utilizes field map(s) and/or equipment parameters to generate guidance lines for the vehicle 102 , which are subsequently employed by the implement commander 220 and the drive commander 222 to enable automated navigation and operation of the vehicle 102 .
  • the operation performance tracker 224 monitors implement operation data to update field maps based on operations that have been performed.
  • FIG. 2 While an example manner of implementing the controller 116 of FIG. 1 is illustrated in FIG. 2 , one or more of the elements, processes and/or devices illustrated in FIG. 2 may be combined, divided, re-arranged, omitted, eliminated and/or implemented in any other way.
  • the example field map handler 202 , the example location data analyzer 204 , the example field map generator 206 , the example map data store 208 , the example field map analyzer 210 , the example user command accessor 212 , the example user interface configurator 214 , the example equipment parameter determiner 216 , the example guidance line generator 218 , the example implement commander 220 , the example drive commander 222 , the example operation performance tracker 224 and/or, more generally, the example controller 116 of FIG. 2 may be implemented by hardware, software, firmware and/or any combination of hardware, software and/or firmware.
  • programmable logic device(s) PLD(s)
  • FPLD field programmable logic device
  • the example controller 116 of FIG. 3 may include one or more elements, processes and/or devices in addition to, or instead of, those illustrated in FIG. 2 , and/or may include more than one of any or all of the illustrated elements, processes and devices.
  • the phrase “in communication,” including variations thereof, encompasses direct communication and/or indirect communication through one or more intermediary components, and does not require direct physical (e.g., wired) communication and/or constant communication, but rather additionally includes selective communication at periodic intervals, scheduled intervals, aperiodic intervals, and/or one-time events.
  • FIG. 3 is a schematic illustrating an example coverage-based mapping technique and an example row-based mapping technique.
  • the schematic includes an example coverage map 302 and an example row map 304 .
  • the example coverage map 302 includes shading on areas where an operation has been performed. In the example coverage map 302 , rectangular areas are utilized as approximations of where the implement 108 performed an operation.
  • the guidance line generator 218 would not be able to identify locations of specific crop rows, as the coverage map 302 merely indicates generally where an operation was performed.
  • the example row map 304 of the illustrated example of FIG. 3 includes example crop row lines 308 a - d , which illustrate precise locations of crop rows on the field.
  • the row map 304 further includes an example guidance line 310 , illustrating a path that the vehicle 102 followed or is to follow while traversing the field.
  • the crop row lines 308 a - d can be integrated in the field map by utilizing data from the operation performance tracker 224 , the equipment parameter determiner 216 , and/or the location data analyzer 204 .
  • the location data analyzer 204 can determine a location of the vehicle 102 based on location data from the location sensor 114 .
  • Accurate positions of individual portions of the implement 108 can be determined by extrapolating the location data for the vehicle 102 based on a known width of the implement and a known location of the location sensor 114 relative to the implement 108 .
  • the operation performance tracker 224 can determine whether portions of the implement 108 were actually utilized during an operation, which therefore determines whether those crop rows should exist in a new field map. For example, if a map is generated during an initial planting operation, the crop row lines 308 a - d should only be generated at locations of crop rows that are actually planted, according to the operation performance tracker 224 .
  • Utilizing a row-based mapping technique allows subsequent generation of accurate guidance lines to perform an operation upon the crop rows using a different implement (which may have a different width) at a later time. Conversely, utilizing a coverage-based map which only depicts general areas which have been operated upon may result in guidance lines being generated that cause overlap of operations, damage to crops, and/or crop rows remaining untreated during subsequent operations.
  • FIG. 4A is an example field map 402 generated using an example four-row implement.
  • the field map 402 of the illustrated example of FIG. 4A includes a plurality of example crop row lines 404 and a plurality of example original guidance lines 406 .
  • the crop row lines 404 accurately illustrate locations of crop rows on the field.
  • the crop row lines 404 can accurately represent crop row locations based on location data determined directly on the implement, therefore accounting for potential implement drift during the operation.
  • the original guidance lines 4 depict a path that was followed by the vehicle which generated the field map.
  • For each one of the original guidance lines 406 there are four crop row lines 404 distributed around the one of the original guidance lines 406 , indicating that the implement of the original vehicle that performed the mapping had an operational width of four-rows. While only one of the crop row lines 404 and one of the original guidance lines 406 are labeled, all lines in FIGS. 4A-4F which have the same line type (using the same dashing pattern) are understood to refer to the same type of element.
  • FIG. 4B is an example first field map traversal schematic 408 using an example twelve-row sprayer 410 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A .
  • the first field map traversal schematic 408 includes an example guidance line 412 for the twelve-row sprayer 410 .
  • the twelve-row sprayer 410 has an operational width that is three times larger than the operational width of the implement which performed the original operation. Thus, there is only the guidance line 412 in the first field map traversal schematic 408 in order for the twelve-row sprayer to perform a spraying operation on twelve of the crop row lines 404 .
  • FIG. 4C is an example second field map traversal schematic 414 using an example eight-row sprayer 416 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A .
  • the second field map traversal schematic 414 includes an example guidance line 418 for the eight-row sprayer 416 .
  • the eight-row sprayer 416 has an operational width that is two times larger than the operational width of the implement which performed the original operation. Thus, there is only the guidance line 418 in the second field map traversal schematic 414 in order for the eight-row sprayer 416 to perform a spraying operation on eight of the crop row lines 404 .
  • FIG. 4D is an example third field map traversal schematic 420 using an example four-row harvester 422 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A .
  • the third field map traversal schematic 420 includes example guidance lines 424 for the four-row harvester 422 . While the guidance lines 424 of the third field map traversal schematic 420 substantially overlap the original guidance lines 406 , in some examples, due to implement drift or other factors, the actual crop row locations may be different such that the guidance lines 424 are different from the original guidance lines 406 to best operate upon the crop rows, despite the operational width of the four-row harvester 422 being the same as the four-row implement of the original operation.
  • FIG. 4E is an example fourth field map traversal schematic 426 using an example two-row harvester 428 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A .
  • the fourth field map traversal schematic 426 includes example guidance lines 430 for the two-row harvester 428 . Since the two-row harvester 428 has an operational width that is half that of the four-row implement used in the original operation, there are twice as many of the guidance lines 430 in the fourth field map traversal schematic 426 relative to the original guidance lines 406 .
  • FIG. 4F is an example fifth field map traversal schematic 432 using an example one-row harvester 434 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A .
  • the fifth field map traversal schematic 432 includes example guidance lines 436 which are displayed directly over the crop row lines 404 of the field map 402 , since the one-row harvester 434 must individually pass over each crop row to harvest the field.
  • the one-row harvester 434 therefore requires four times as many guidance lines as the four-row implement used in the original operation.
  • FIG. 5A is an enlarged view of the example field map 402 of FIG. 4A generated using a four-row implement.
  • FIG. 5B is an example sixth field map traversal schematic 502 using an example single row harvester 504 with an example haulage vehicle 506 on a portion of the field corresponding to the field map 402 of FIG. 5A .
  • the single row harvester 504 is a first implement which harvests the crop rows
  • the haulage vehicle 506 is a second implement which is used to store the harvest during operation.
  • the controller 116 must generate guidance lines for the vehicle based on operational widths and relative positioning of both the single row harvester 504 and the haulage vehicle 506 .
  • the guidance lines for the single row harvester 504 and the haulage vehicle 506 are omitted from FIG. 5B , and instead illustrated with each of the implements individually in FIGS. 5C and 5D .
  • the controller e.g., the guidance line generator 218
  • the controller may generate a single guidance line for the overall vehicle to follow based on the individual guidance lines for the implements.
  • FIG. 5C is an example seventh field map traversal schematic 508 using the example single row harvester 504 of FIG. 5B on the field corresponding to the field map 402 of FIG. 5A .
  • example guidance lines 510 of FIG. 5C for the single row harvester 504 overlap on all crop rows of the field map 402 to ensure all crop rows are operated upon. While the guidance lines 510 of the illustrated example substantially overlap the original guidance lines 406 , in some examples, even when the operational width of the vehicle implement performing the subsequent operation is the same as the operational width of the vehicle implement performing the original operation, the guidance lines may not overlap. For example, due to drift of the implement, crop row locations may not be evenly distributed around the original guidance line.
  • the guidance lines may be generated in different locations than the original guidance lines to better target the actual locations of the crop rows.
  • the benefit of having knowledge of specific locations of the crop rows can be realized even in examples where a subsequent vehicle has a same operational width as an original vehicle.
  • FIG. 5D is an example eighth field map traversal schematic 512 using the example haulage vehicle 506 of FIG. 5B on the field corresponding to the field map 402 of FIG. 5A .
  • the eighth field map traversal schematic 512 includes example guidance lines 514 for the haulage vehicle 506 .
  • the guidance lines 514 of the illustrated example of FIG. 5D are generated (e.g., by the guidance line generator 218 ) such that wheels of the haulage vehicle 506 do not drive over the crop rows represented by the crop row lines 404 .
  • the haulage vehicle 506 does not perform an operation on the crop rows itself, but instead assists the single-row harvester 504 .
  • the guidance lines 514 of FIG. 5B are not generated to enable an operation on the crop rows, but to avoid damaging the crop rows (e.g., by driving over them with the haulage vehicle 506 ).
  • FIG. 6A is an example ninth field map traversal schematic 602 illustrating an example original operation and an example subsequent operation executed on a field map with a narrow guess row.
  • the ninth field map traversal schematic 602 includes an original vehicle 604 which performed the original operation and generate the field map. The two instances of the original vehicle 604 represents subsequent passes of the same original vehicle 604 .
  • the ninth field map traversal schematic 602 includes example crop row lines 606 indicating locations at which the crop rows were operated upon (e.g., planted) during the original operation.
  • the original operation exhibited an example narrow guess row 608 , meaning that outer crop rows from consecutive passes of the original vehicle 604 have a narrow spacing relative to other crop rows. Without accurate knowledge of the crop row locations (e.g., as provided by field maps generated using techniques disclosed herein), subsequent operations performed on the field may drive over the narrow guess row 608 , and/or otherwise improperly operate upon the crop rows of the narrow guess row 608 .
  • the ninth field map traversal schematic 602 includes an example subsequent vehicle 610 with a twelve-row implement.
  • the twelve-row implement could theoretically operate upon all of the crop rows represented by the crop row lines 606 , but to do so, the subsequent vehicle 610 would have to drive in the center of the crop rows, and may end up driving over the narrow guess row 608 .
  • example guidance lines 612 are generated accounting for the specific positions of the crop rows and the width of the implement of the subsequent vehicle 610 .
  • the example guidance lines 612 cause the subsequent vehicle 610 to travel between rows with regular widths, thus avoiding driving through the narrow guess row 608 .
  • the implement commander 220 of the controller 116 can cause portions of the implement to be enabled and/or disabled based on whether crop rows beneath the implement have been operated upon. Based on the positioning of the guidance lines 612 , utilizing all portions of the implement throughout the entire traversal would result in repeated operations, and hence, the implement must be selectively activated over rows which have not yet been operated upon.
  • FIG. 6B is an example tenth field map traversal schematic 614 illustrating another original operation and another subsequent operation executed on a field map with an accurate guess row.
  • the tenth field map traversal schematic 614 includes the example original vehicle 604 and the example subsequent vehicle 610 .
  • the original operation is performed by the original vehicle 604 with example crop row lines 606 including an example accurate guess row 618 (e.g., a guess row where the spacing between outer crop rows operated upon by the original implement during consecutive passes are spaced approximately equally to other crop rows).
  • the subsequent vehicle 610 utilizes an example guidance line 620 that is in between crop rows of the accurate guess row 618 .
  • the guidance line generator 218 of the illustrated example of FIG. 2 can generate the guidance line 620 centrally such that the twelve-row implement of the subsequent vehicle 610 can operate upon all of the twelve crops rows without the subsequent vehicle 610 running over any of the crop rows (e.g., as may happen utilizing the guidance line 620 of FIG. 6B with the narrow guess row of FIG. 6A .
  • guidance lines generated for automated vehicle operation utilizing techniques disclosed herein enable accurate and efficient operation on afield regardless of unique spacing of crop rows (e.g., regardless of differently-sized guess rows, regardless of contours of the field, etc.).
  • FIG. 7 is an example run screen 702 to be displayed on the example user interface 122 of the vehicle 102 of FIG. 1 .
  • the example run screen 702 includes an example navigation display portion 704 to display mapping data, guidance lines, crop row locations, and/or other information pertinent to navigation of the vehicle 102 .
  • the navigation display portion 704 includes an example vehicle position indicator 706 displaying a current position of the vehicle 102 relative to the field illustrated in the navigation display portion 704 .
  • the navigation display portion 704 further includes example crop row lines 708 , illustrating precise locations of crop rows, as determined from a previously generated field map (e.g., generated by the field map generator 206 during an initial operation).
  • a previously generated field map e.g., generated by the field map generator 206 during an initial operation.
  • the navigation display portion 704 includes an example current guidance line 710 (displayed as a solid line) that the vehicle 102 is following during its automated operation on the field.
  • the current guidance line 710 is generated by the guidance line generator 218 of the controller 116 of FIG. 2 .
  • the navigation display portion 704 includes example past or future guidance lines 712 , which indicate guidance lines that are either not yet being followed or were already followed during previous operation.
  • the navigation display portion 704 includes an example implement position line 714 , indicating a current position and operational width of the implement 108 .
  • Behind the implement (e.g., under the implement, in the view of FIG. 7 ) on the field is an example coverage section 716 .
  • the coverage section 716 provides shading for areas which have been operated upon by the implement 108 .
  • the operation performance tracker 224 indicates portions of the implement 108 which have been enabled throughout an operation. In some such examples, this implement operation data can be combined with location data indicating a position of the implement 108 to plot the coverage section 716 .
  • the run screen 702 of the illustrated example of FIG. 7 includes an example guidance portion 718 .
  • the guidance portion 718 includes an example set track button 720 to allow an operate to determine a track (e.g., one or more guidance lines) for the vehicle 102 to follow.
  • the tracks e.g., the plurality of guidance lines
  • the set track button 720 allows the user to select a guidance line and/or begin the process of manually creating a new guidance line (e.g., in the absence of mapping data).
  • the guidance portion 718 further includes an example track adjustment portion 722 to enable manual adjustments of the guidance line. For example, if the current guidance line 710 appears to be problematic (e.g., the operator notices that the vehicle 102 is driving over a crop row), the operator can manually shift a position of the current guidance line 710 .
  • the run screen 702 of the illustrated example of FIG. 7 further includes an example setup button 724 .
  • the setup button 724 when pressed, allows an operator to adjust parameters of the vehicle 102 and/or the implement 108 . For example, if the implement 108 does not automatically communicate parameters (e.g., an operational width, an implement type, etc.) to the controller 116 , the operator can enter these parameters into a prompt provided when the setup button 724 is pressed.
  • parameters e.g., an operational width, an implement type, etc.
  • the run screen 702 of the illustrated example of FIG. 7 includes an example work recording button 726 .
  • the controller 116 When actuated to “on,” as shown in the illustrated example of FIG. 7 , the controller 116 records implement operation data, thereby enabling generation of the coverage section 716 .
  • the operation performance tracker 224 communicates operation performance data to the map data store 208 , the field map analyzer 210 , and/or the field map generator 206 , such that the coverage section 716 can be stored in the field map and accessed by an operator to view portions of the field which have been operated upon.
  • the run screen 702 of the illustrated example of FIG. 7 includes an example auto-trac button 728 , which allows an operator to enable or disable automated guidance of the vehicle 102 and/or automated operation of the implement 108 . Since the auto-trac button is listed as enabled in FIG. 7 , the vehicle 102 is operating in an automated mode, automatically generating and following guidance lines and performing operations with the implement 108 based on existing field maps.
  • FIGS. 8-10 Flowcharts representative of example hardware logic, machine readable instructions, hardware implemented state machines, and/or any combination thereof for implementing the controller 116 of FIGS. 1 and 2 are shown in FIGS. 8-10 .
  • the machine readable instructions may be one or more executable programs or portion(s) of an executable program for execution by a computer processor such as the processor 1112 shown in the example processor platform 1100 discussed below in connection with FIG. 11 .
  • the program may be embodied in software stored on a non-transitory computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a DVD, a Blu-ray disk, or a memory associated with the processor 1112 , but the entire program and/or parts thereof could alternatively be executed by a device other than the processor 1112 and/or embodied in firmware or dedicated hardware.
  • a non-transitory computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a DVD, a Blu-ray disk, or a memory associated with the processor 1112 , but the entire program and/or parts thereof could alternatively be executed by a device other than the processor 1112 and/or embodied in firmware or dedicated hardware.
  • FIGS. 8-10 many other methods of implementing the example controller 116 may alternatively be used. For example, the order of execution of the blocks may be changed, and/or some of the blocks described may be changed, eliminated, or combined.
  • any or all of the blocks may be implemented by one or more hardware circuits (e.g., discrete and/or integrated analog and/or digital circuitry, an FPGA, an ASIC, a comparator, an operational-amplifier (op-amp), a logic circuit, etc.) structured to perform the corresponding operation without executing software or firmware.
  • hardware circuits e.g., discrete and/or integrated analog and/or digital circuitry, an FPGA, an ASIC, a comparator, an operational-amplifier (op-amp), a logic circuit, etc.
  • the machine readable instructions described herein may be stored in one or more of a compressed format, an encrypted format, a fragmented format, a packaged format, etc.
  • Machine readable instructions as described herein may be stored as data (e.g., portions of instructions, code, representations of code, etc.) that may be utilized to create, manufacture, and/or produce machine executable instructions.
  • the machine readable instructions may be fragmented and stored on one or more storage devices and/or computing devices (e.g., servers).
  • the machine readable instructions may require one or more of installation, modification, adaptation, updating, combining, supplementing, configuring, decryption, decompression, unpacking, distribution, reassignment, etc.
  • the machine readable instructions may be stored in multiple parts, which are individually compressed, encrypted, and stored on separate computing devices, wherein the parts when decrypted, decompressed, and combined form a set of executable instructions that implement a program such as that described herein.
  • the machine readable instructions may be stored in a state in which they may be read by a computer, but require addition of a library (e.g., a dynamic link library (DLL)), a software development kit (SDK), an application programming interface (API), etc. in order to execute the instructions on a particular computing device or other device.
  • a library e.g., a dynamic link library (DLL)
  • SDK software development kit
  • API application programming interface
  • the machine readable instructions may need to be configured (e.g., settings stored, data input, network addresses recorded, etc.) before the machine readable instructions and/or the corresponding program(s) can be executed in whole or in part.
  • the disclosed machine readable instructions and/or corresponding program(s) are intended to encompass such machine readable instructions and/or program(s) regardless of the particular format or state of the machine readable instructions and/or program(s) when stored or otherwise at rest or in transit.
  • FIGS. 8-10 may be implemented using executable instructions (e.g., computer and/or machine readable instructions) stored on a non-transitory computer and/or machine readable medium such as a hard disk drive, a flash memory, a read-only memory, a compact disk, a digital versatile disk, a cache, a random-access memory and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information).
  • a non-transitory computer readable medium is expressly defined to include any type of computer readable storage device and/or storage disk and to exclude propagating signals and to exclude transmission media.
  • A, B, and/or C refers to any combination or subset of A, B, C such as (1) A alone, (2) B alone, (3) C alone, (4) A with B. (5) A with C. (6) B with C, and (7) A with B and with C.
  • the phrase “at least one of A and B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B. and (3) at least one A and at least one B.
  • the phrase “at least one of A or B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B, and (3) at least one A and at least one B.
  • the phrase “at least one of A and B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B, and (3) at least one A and at least one B.
  • the phrase “at least one of A or B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B and (3) at least one A and at least one B.
  • Example machine readable instructions 800 that may be executed by the controller 116 to perform an operation on a field using a field map are illustrated in FIG. 8 .
  • the example machine readable instructions 800 begin with the example controller 116 determining equipment parameter(s) automatically or via manual entry (Block 802 ).
  • the equipment parameter determiner 216 determines equipment parameter(s) automatically (e.g., based on data from the implement 108 and/or from the implement sensors 112 a - 1 ), and/or via manual entry (e.g., via entry from the operator on the user interface 122 ).
  • the example controller 116 determines if the current operation is the first operation on the field.
  • the field map analyzer 210 determines if the current operation is the first operation on the field by comparing a location of the vehicle 102 , as determined by the location data analyzer 204 , with existing field maps in the map data store 208 . If a field map exists, then the current operation is not the first operation on the field. Conversely, if a field map does not exist, the current operation is the first operation on the field. In response to the current operation being the first operation on the field, processing transfers to block 806 . Conversely, in response to the current operation not being the first operation on the field, processing transfers to block 808 .
  • the example controller 116 generates a field map during a first operation. Further detail of the implementation of block 806 is illustrated and described below in connection with FIG. 9 .
  • the example controller 116 retrieves a map associated with the field.
  • the field map analyzer 210 retrieves a field map from the map data store 208 and/or form another location (e.g., a network location) accessible to the controller 116 .
  • the example controller 116 determines crop row locations.
  • the field map analyzer 210 determines crop row locations based on the field map.
  • the field map analyzer 210 determines crop row locations based on the field map and historical implement operation data, as analyzed by the operation performance tracker 224 .
  • the example controller 116 determines guidance line(s) based on crop row locations and implement parameters.
  • the guidance line generator 218 generates guidance lines based on crop row locations from the field map analyzer 210 and implement parameters from the equipment parameter determiner 216 .
  • the guidance line generator 218 can utilize a known operational width of the implement 108 to generate guidance lines to perform an operation upon the crop rows.
  • the example controller 116 displays mapping data and a guidance line on the user interface 122 .
  • the user interface configurator causes the user interface 122 to display mapping data (e.g., a field map) and a guidance line on the user interface 122 .
  • mapping data e.g., a field map
  • An example of such mapping data and a guidance line displayed on the user interface 122 is illustrated by the run screen 702 of FIG. 7 .
  • the example controller 116 automatically guides the vehicle 102 along guidance line(s) and perform operation(s). Further detail of the implementation of block 816 is illustrated and described below in connection with FIG. 10 .
  • the example controller 116 determines whether the operation is complete.
  • the user command accessor 212 determines, based on user commands entered on the user interface 122 , whether the operation is complete.
  • the operation performance tracker 224 determines, based on implementation operation data and a location of the vehicle 102 , whether the operation is complete. In response to the operation being complete, processing transfers to block 820 . Conversely, in response to the operation not being complete, processing transfers to block 814 .
  • the example controller 116 determines whether there is an additional operation to perform.
  • the user command accessor 212 determines, based on user commands, whether there is an additional operation to perform on the field. In response to there being an additional operation to perform, processing transfers to block 802 . Conversely, in response to there not being an additional operation to perform, processing transfers to block 822 .
  • the example controller 116 saves implement operation data in association with the field map.
  • the operation performance tracker 242 , the field map analyzer 210 , the field map generator 206 , and/or the map data store 208 save implement operation data in association with the field map.
  • Example machine readable instructions 900 that may be executed by the controller 116 to generate a first operation map are illustrated in FIG. 9 .
  • the example machine readable instructions 900 begin with the example controller 116 determining equipment parameter(s) automatically or via manual entry.
  • the equipment parameter determiner 216 determines equipment parameter(s) automatically (e.g., based on data from the implement 108 and/or from the implement sensors 112 a - l ), and/or via manual entry (e.g., via entry from the operator on the user interface 122 ).
  • the example controller 116 begins manual or automated motion of the vehicle 102 across the field.
  • the operator controls motion of the vehicle 102 by commanding the drive commander 222 , and controls usage of the implement 108 by commanding the implement commander 220 .
  • the controller 116 automatically operates the vehicle via the drive commander 222 and the implement commander 220 by generating general guidance lines (e.g., evenly spaced, linear guidance lines) based on a field dimension and continually operating upon the field with all portions of the implement 108 via the implement commander 220 .
  • general guidance lines e.g., evenly spaced, linear guidance lines
  • the example controller 116 collects location data during operation.
  • the location data analyzer 204 collects GPS data and inertial data from the location sensor 114 during operation of the vehicle 102 .
  • the example controller 116 determines if the field operation is complete.
  • the user command accessor 212 determines, based on user commands entered into the user interface 122 , whether the field operation is complete. In response to the field operation being complete, processing transfers to block 910 . Conversely, in response to the field operation not being complete, processing transfers to block 906 .
  • the example controller 116 generates a map for the field which has been operated upon.
  • the field map generator 206 generates a map for the field.
  • the field map generator 206 can generate the map based on location data from the location data analyzer 204 and/or implement operation data from the operation performance tracker 224 .
  • the example controller 116 stores the field map and/or uploads the field map to a network and/or central computing system.
  • the map data store 208 is used to store the field map.
  • Example machine readable instructions 100 that may be executed by the controller 116 to automatically guide a vehicle along guidance line(s) and perform an operation on a field are illustrated in FIG. 10 .
  • the example machine readable instructions 1000 begin with the example controller 116 advancing the vehicle 102 along a guidance line accounting for curvature and/or contour of the field (Block 1002 ).
  • the drive commander 222 causes the vehicle 102 to advance along the guidance line through the field.
  • the example controller 116 activates portions of the implement 108 to perform operations on rows which have not been operated upon.
  • the implement commander 220 activates (e.g., enables, causes use of, etc.) portions of the implement 108 to perform operations on crop rows which have not yet been operated upon, based on implement operation data as analyzed by the operation performance tracker 224 and based on the field map, as analyzed by the field map analyzer 210 .
  • the example controller 116 records implement operation data.
  • the operation performance tracker 224 records implement operation data by storing it in association with a field map of the current field being operated upon in the map data store 208 .
  • FIG. 11 is a block diagram of an example processor platform 1100 structured to execute the instructions of FIGS. 8-10 to implement the controller 116 of FIGS. 1 and 2 .
  • the processor platform 1100 can be, for example, a server, a personal computer, a workstation, a self-learning machine (e.g., a neural network), a mobile device (e.g., a cell phone, a smart phone, a tablet such as an iPadTM), a personal digital assistant (PDA), an Internet appliance, a DVD player, a CD player, a digital video recorder, a Blu-ray player, a gaming console, a personal video recorder, a set top box, a headset or other wearable device, or any other type of computing device.
  • a self-learning machine e.g., a neural network
  • a mobile device e.g., a cell phone, a smart phone, a tablet such as an iPadTM
  • PDA personal digital assistant
  • an Internet appliance e.g., a
  • the processor platform 1100 of the illustrated example includes a processor 1112 .
  • the processor 1112 of the illustrated example is hardware.
  • the processor 1112 can be implemented by one or more integrated circuits, logic circuits, microprocessors, GPUs, DSPs, or controllers from any desired family or manufacturer.
  • the hardware processor may be a semiconductor based (e.g., silicon based) device.
  • the processor implements the example field map handler 202 , the example location data analyzer 204 , the example field map generator 206 , the example map data store 208 , the example field map analyzer 210 , the example user command accessor 212 , the example user interface configurator 214 , the example equipment parameter determiner 216 , the example guidance line generator 218 , the example implement commander 220 , the example drive commander 222 , and the example operation performance tracker 224 .
  • the processor 1112 of the illustrated example includes a local memory 1113 (e.g., a cache).
  • the processor 1112 of the illustrated example is in communication with a main memory including a volatile memory 1114 and a non-volatile memory 1116 via a bus 1118 .
  • the volatile memory 1114 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS® Dynamic Random Access Memory (RDRAM®) and/or any other type of random access memory device.
  • the non-volatile memory 1116 may be implemented by flash memory and/or any other desired type of memory device. Access to the main memory 1114 , 1116 is controlled by a memory controller.
  • the processor platform 1100 of the illustrated example also includes an interface circuit 1120 .
  • the interface circuit 1120 may be implemented by any type of interface standard, such as an Ethernet interface, a universal serial bus (USB), a Bluetooth® interface, a near field communication (NFC) interface, and/or a PCI express interface.
  • one or more input devices 1122 are connected to the interface circuit 1120 .
  • the input device(s) 1122 permit(s) a user to enter data and/or commands into the processor 1012 .
  • the input device(s) can be implemented by, for example, an audio sensor, a microphone, a camera (still or video), a keyboard, a button, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system.
  • One or more output devices 1124 are also connected to the interface circuit 1120 of the illustrated example.
  • the output devices 1024 can be implemented, for example, by display devices (e.g., a light emitting diode (LED), an organic light emitting diode (OLED), a liquid crystal display (LCD), a cathode ray tube display (CRT), an in-place switching (IPS) display, a touchscreen, etc.), a tactile output device, a printer and/or speaker.
  • the interface circuit 1120 of the illustrated example thus, typically includes a graphics driver card, a graphics driver chip and/or a graphics driver processor.
  • the interface circuit 1120 of the illustrated example also includes a communication device such as a transmitter, a receiver, a transceiver, a modem, a residential gateway, a wireless access point, and/or a network interface to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 1126 .
  • the communication can be via, for example, an Ethernet connection, a digital subscriber line (DSL) connection, a telephone line connection, a coaxial cable system, a satellite system, a line-of-site wireless system, a cellular telephone system, etc.
  • DSL digital subscriber line
  • the processor platform 1100 of the illustrated example also includes one or more mass storage devices 1128 for storing software and/or data Examples of such mass storage devices 1128 include floppy disk drives, hard drive disks, compact disk drives, Blu-ray disk drives, redundant array of independent disks (RAID) systems, and digital versatile disk (DVD) drives.
  • mass storage devices 1128 include floppy disk drives, hard drive disks, compact disk drives, Blu-ray disk drives, redundant array of independent disks (RAID) systems, and digital versatile disk (DVD) drives.
  • the machine executable instructions 1132 of FIGS. 1 and 2 may be stored in the mass storage device 1128 , in the volatile memory 1114 , in the non-volatile memory 1116 , and/or on a removable non-transitory computer readable storage medium such as a CD or DVD.
  • example methods, apparatus and articles of manufacture have been disclosed that enable accurate automated operation of an agricultural vehicle to perform operations on a field by utilizing field mapping data to generate guidance lines that account for differences between types of implements, unique topographical features of the field, unique crop row geometries, and/or other factors.

Landscapes

  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Mechanical Engineering (AREA)
  • Soil Sciences (AREA)
  • Environmental Sciences (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Guiding Agricultural Machines (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Methods, apparatus, systems and articles of manufacture are disclosed for field operations based on historical field operation data. An example apparatus disclosed herein includes a guidance line generator to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field. The example apparatus further includes a drive commander to cause the vehicle to traverse the field along the guidance line, and an implement commander to cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.

Description

FIELD OF THE DISCLOSURE
This disclosure relates generally to agricultural techniques, and, more particularly, to apparatus and methods for field operations based on historical field operation data.
BACKGROUND
In recent years, agricultural vehicles have become increasingly automated. Agricultural vehicles may perform semi-autonomous or fully-autonomous operations on fields. Agricultural vehicles perform operations using implements including planting implements, spraying implements, harvesting implements, fertilizing implements, strip/till implements, etc.
SUMMARY
An example apparatus disclosed herein includes a guidance line generator to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field. The example apparatus further includes a drive commander to cause the vehicle to traverse the field along the guidance line and an implement commander to cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
An example non-transitory computer readable storage medium disclosed herein includes instructions that, when executed, cause a processor to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field. The instructions, when executed, further cause the processor to cause the vehicle to traverse the field along the guidance line, and cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
An example method disclosed herein includes generating a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows and (2) an implement of the vehicle, the implement to perform the second operation on the field. The example method further includes causing the vehicle to traverse the field along the guidance line and causing the implement to perform the second operation as the vehicle traverses the field along the guidance line.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a schematic of an example vehicle constructed in accordance with the teachings of this disclosure.
FIG. 2 is a schematic of an example controller of the vehicle of FIG. 1.
FIG. 3 is a schematic illustrating an example coverage-based mapping technique and an example row-based mapping technique.
FIG. 4A is an example field map generated using an example four-row implement.
FIG. 4B is an example first field map traversal schematic using an example twelve-row sprayer in a subsequent operation on the field corresponding to the field map of FIG. 4A.
FIG. 4C is an example second field map traversal schematic using an example eight-row sprayer in a subsequent operation on the field corresponding to the field map of FIG. 4A.
FIG. 4D is an example third field map traversal schematic using an example four-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A.
FIG. 4E is an example fourth field map traversal schematic using an example two-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A.
FIG. 4F is an example fifth field map traversal schematic using an example one-row harvester in a subsequent operation on the field corresponding to the field map of FIG. 4A.
FIG. 5A is an enlarged view of the example field map of FIG. 4A generated using a four-row implement.
FIG. 5B is an example sixth field map traversal schematic using an example single row harvester with an example haulage vehicle on the field corresponding to the field map of FIG. 5A.
FIG. 5C is an example seventh field map traversal schematic using the example single row harvester of FIG. 5B on the field corresponding to the field map of FIG. 5A.
FIG. 5D is an example eighth field map traversal schematic using the example haulage vehicle of FIG. 5B on the field corresponding to the field map of FIG. 5A.
FIG. 6A is an example ninth field map traversal schematic illustrating an original operation and a subsequent operation executed on a field map with a narrow guess row.
FIG. 6B is an example tenth field map traversal schematic illustrating another original operation and another subsequent operation executed on a field map with an accurate guess row.
FIG. 7 is an example run screen to be displayed on an example user interface of the vehicle of FIG. 1.
FIG. 8 is a flowchart representative of example machine readable instructions that may be executed to implement the controller of FIGS. 1 and 2 to perform an operation on a field using a field map.
FIG. 9 is a flowchart representative of example machine readable instructions that may be executed to implement the controller of FIGS. 1 and 2 to generate a first operation map.
FIG. 10 is a flowchart representative of example machine readable instructions that may be executed to implement the controllers of FIGS. 1 and 2 to automatically guide a vehicle along guidance line(s) and perform an operation on a field.
FIG. 11 is a block diagram of an example processing platform structured to execute the example machine readable instructions of FIGS. 8-10 to implement the controller of FIGS. 1 and 2. The figures are not to scale. In general, the same reference numbers will be used throughout the drawing(s) and accompanying written description to refer to the same or like parts.
Descriptors “first,” “second,” “third,” etc. are used herein when identifying multiple elements or components which may be referred to separately. Unless otherwise specified or understood based on their context of use, such descriptors are not intended to impute any meaning of priority, physical order or arrangement in a list, or ordering in time but are merely used as labels for referring to multiple elements or components separately for ease of understanding the disclosed examples. In some examples, the descriptor “first” may be used to refer to an element in the detailed description, while the same element may be referred to in a claim with a different descriptor such as “second” or “third.” In such instances, it should be understood that such descriptors are used merely for ease of referencing multiple elements or components.
DETAILED DESCRIPTION
Agricultural vehicles perform numerous operations on fields using different implements. As an example, a planter implement may plant crops at the beginning of a season. A sprayer implement may subsequently spray the crops, and/or a fertilizer implement may subsequently fertilize the crops. Different vehicles may be used for individual operations, or different implements may be attached to a same vehicle (e.g., a tractor) to perform different operations. In some examples, a planting operation, a strip/till operation, or a fertilizer operation may be performed first, followed by a spraying operation, a harvesting operation (e.g., a combine or forage harvester), a side/top dress fertilizing operation, or any other field operation. As used herein, an “operation” can include planting, spraying, fertilizing, and/or performing any other task on a field. In some examples, operations performed on the field are additionally or alternatively referred to as “treatments.” Further, performing such operations on a field can be referred to as “operating upon” the field, “treating” the field, or “working” the field. In some examples, an operation may include installation of permanent assets, such as beds or drip irrigation lines. While example orderings of operations (e.g., planting followed by fertilizing) may be discussed herein, operations may be performed in any order.
Automation of agricultural vehicles is highly commercially desirable, as automation can improve the accuracy with which operations are performed, reduce operator fatigue, improve efficiency, and accrue other benefits. Automated vehicles move by following guidance lines. Conventional guidance lines are generated by defining a first point (e.g., an “A” point) and a second point (e.g., a “B” point) and generating a line or curve connecting the first point and the second point (e.g., an A-B line, or an A-B curve). Automatically generating point-to-point guidance lines across a field dimension fails to account for topography of the field and specific locations of crop rows. Most operations are subsequent operations performed on a field (e.g., most operations are performed on existing crops that are planted in the field).
In conventional automated agricultural vehicle systems, subsequent operations may be manually performed by an operator or automatically performed while accepting damage to the crop rows. In manual operations, the operator analyzes the crop rows and attempts to position the agricultural vehicle to operate on the crop rows to minimize damage to crops or waste of product (e.g., spray, fertilizer, etc.). In some conventional examples, subsequent field operations are performed automatically based on planned or expected crop row locations. However, due to sources of error, such as estimation of implement drift in curved paths, drift due to slopes in the terrain, uneven draft loads across the width of an implement which cause the implement to pull off-center, disturbances due to soil surface roughness, and/or other sources of error, planned crop row locations are often significantly different than actual crop row locations.
In some conventional implementations, agricultural vehicle systems attempt to sense locations of crop rows as they operate. However, such real-time sensing is insufficient to avoid crop damage, and fails to reliably address unique crop row locations. Further, to ensure all crops are operated upon (sprayed, fertilized, etc.), conventional agricultural vehicle systems may utilize an overlap, where crop rows at the edge of the implement get operated upon twice in two successive passes, resulting in inefficiency and waste.
When a planting operation is performed, after completing planting of one set of rows (e.g., one-pass of the field in a direction), the planter must return in a new pass in the opposite direction. To begin a new set of rows, the operator must estimate a starting position for the first row at an end of the planting implement. Specifically, the operator must try and line up the vehicle such that the first new row is spaced apart from the nearest existing crop row at the same spacing as exists between the rest of the rows. Such edge rows are referred to as “guess rows,” since the operator must guess the appropriate spacing to plant the crop row. This is a very challenging task, which often results in the first row having too narrow a gap relative to the nearest existing row, or too large a gap. In subsequent operations, due to the irregularity of guess rows, an operator may have difficulty performing an operation on the guess rows. Additionally, the size of the guess row can vary from pass to pass (e.g., between traversals) over the field, introducing additional variability that makes subsequent operations increasingly difficult. In some conventional implementations, an operator will merely drive over and sacrifice these “guess rows” during subsequent operations.
Further, in conventional agricultural systems, differences between operational widths of implements complicate determination of driving paths. For example, if a twelve-row planter is utilized to plant the crops, following the same guidance lines used with the twelve-row planter when using a sixteen-row cultivator/side-dress rig for a fertilization operation will result in crop rows that are not sprayed. Farmers will often have a plurality of implements with different widths, and thus operators must manually determine driving paths that are appropriate based on the unique crop rows on the field and unique implement widths. Further, even when the operational width of an implement in a subsequent operation is different than the operational width of the implement in an original operation, utilizing the same original guidance lines in the subsequent operation fails to account for the actual crop row locations (e.g., the crop rows may not be predictably spaced around the original guidance line due to implement drift).
Example methods, apparatus, systems, and articles of manufacture (e.g., physical storage media) disclosed herein generate guidance lines for automatic execution of subsequent operations on a field based on field maps determined during previous operations. Some examples disclosed herein generate field maps based on location data (e.g., GPS and inertial data) and sensor data from implements to accurately represent crop rows in field maps. Example methods, apparatus, systems, and articles of manufacture (e.g., physical storage media) disclosed herein enable accurate navigation along guidance lines to accurately operate upon a field regardless of the terrain of the field and/or the size of implement utilized. Examples disclosed herein accurately operate upon guess rows by generating guidance lines specific to locations of the guess rows, eliminating potential operator error.
Performing subsequent operations based on field maps improves alignment relative to established row locations and minimizes damage to crops, as the wheels of the agricultural vehicle can be positioned to avoid running over rows. Further, utilizing the automated guidance techniques disclosed herein, fields can be operated upon (e.g., harvested, sprayed, fertilized, etc.) faster since the guidance system can anticipate upcoming changes in the terrain and/or changes in crop row locations, rather than relying upon an operator to make adjustments and/or a sensor to sense changes in the terrain. Additionally, some crops, such as corn, require highly precise positioning of the implement (e.g., a harvesting head, a platform for small grain harvesting, etc.) during harvesting. Utilization of accurate mapping data to generate precise guidance lines helps ensure that crops are operated upon safely and efficiently.
Example techniques disclosed herein enable an operator to begin a subsequent operation at any location in a field, as mapping data can be utilized to automatically generate guidance lines for any location. Further, example techniques disclosed herein enable a coordinated work system in which a plurality of agricultural vehicles complete an operation by each generating guidance lines based off shared mapping data and knowledge of other vehicles' planned operations. Similarly, a plurality of agricultural vehicles can generate the field map during an initial operation (e.g., a planting operation).
Example techniques disclosed herein further utilize individual section and/or nozzle control to ensure that if a portion of the implement is passing over a crop row which has already been operated upon (e.g., already sprayed), this portion of the implement does not perform a redundant operation.
FIG. 1 is a schematic of an example vehicle 102 constructed in accordance with the teachings of this disclosure. The example vehicle 102 includes an example main body 104 including an example cab 106 from which an operator can operate the vehicle 102. In some examples wherein the vehicle 102 is autonomous, the vehicle 102 may not include the cab 106.
The vehicle 102 further includes an example implement 108. In the illustrated example of FIG. 1, the implement 108 is a planting implement. Specifically, the implement 108 is a twelve-row planter. The implement 108 plants crops in example crop rows 110 a-l. The implement 108 may be separable from the vehicle 102, and exchanged with another implement of a same or different implement type. For example, the implement may be a sprayer implement, a fertilizer implement, a tilling implement, a harvesting implement, and/or any other type of implement to perform an operation on a field. In some examples, the implement 108 is not separable from the vehicle 102, and multiple vehicles may be used to perform different operations on a field. While the implement 108 of FIG. 1 is a twelve row planter, the implement 108 may have any operational width (e.g., four rows, eight rows, sixteen rows, etc.). The operational width of the implement 108 defines the number of rows that can be operated upon (e.g., planted, tilled, sprayed, etc.) in a single pass of the vehicle 102 over a field. While the implement 108 of the illustrated example is capable of planting twelve rows simultaneously, in some examples, only some of the rows are planted. For example, if a right-side of the implement 108 is traveling over a portion of a field that has already been planted, the corresponding rows of the implement 108 may be disabled to prevent performing a repeat operation.
Individual row portions of the implement 108 of the illustrated example of FIG. 1 include example implement sensors 112 a-l to detect whether the row portions of the implement 108 are performing operations. While only the implement sensors 112 a-e are visible due to the angle of the perspective view of FIG. 1, additional implement sensors 112 f-l are understood to be mounted to portions of the implement 108 in alignment with crop rows 110 f-l. The implement sensors 112 a-1 are each positioned on one of the portions of the implement 108 to detect whether operations are performed on individual rows. In some examples, the implement sensors 112 a-1 are switches which activate as a product passes through the implement portion (e.g., as seed exits the implement 108). Any type of sensor may be utilized to determine whether a portion of the implement 108 corresponding to a specific row is performing an operation. The implement sensors 112 a-1 communicate with an example controller 116.
The implement 108 of the illustrated example of FIG. 1 includes an example location sensor 114. The location sensor 114 of the illustrated example is a global positioning system (GPS) sensor combined with an inertial sensor. The inertial sensor enables higher precision location data by enabling compensation for tilt of the vehicle 102. For example, if the vehicle 102 is tilted on a hill, the inertial sensor can compensate for this tilt to enable accurate determination of a location of the vehicle 102, and more specifically, accurate determination of locations of portions of the implement 108. The example location sensor 114 of the illustrated example is mounted to the implement 108 to provide more accurate ground-level location data for subsequent determination of crop row locations. In some examples, the location sensor 114 is mounted to another part of the vehicle 102 (e.g., the main body 104, the cab 106, etc.). In some examples, the location sensor 114 includes multiple location sensors. In some such examples, the vehicle 102 includes the location sensor 114 as a first sensor on the main body 104 to provide a location for the vehicle 102 generally, and a second sensor on the implement 108 to provide ground-level location data. The location sensor 114 communicates to an example controller 116.
The example controller 116 of the illustrated example of FIG. 1 analyzes data from one or more sensors and enables automated operation of the vehicle 102. The controller 116 of the illustrated example generates field maps and utilizes such field maps to automatically generate guidance lines based on the field maps and the type of implement in use by the vehicle 102. The controller 116 accesses implement sensor data from the implement sensors 112 a-l and location data from the location sensor 114 to generate field maps. For example, the controller 116 can determine locations of the crop rows 110 a-l based on the implement sensor data and the location data. The controller 116 of the illustrated example issues commands to cause movement of the vehicle 102 by commanding an example drive controller 118. Further, the controller 116 of the illustrated example issues commands to cause operation of the implement 108 by commanding an example implement controller 120 to perform an operation based on a field map and a parameter of the implement 108. The controller 116 of the illustrated example further controls an example user interface 122, which can provide mapping data to an operator and solicit input from the operator. Further detail pertaining to the controller 116 is illustrated and described in association with FIG. 2.
The example drive controller 118 of the illustrated example of FIG. 1 causes movement of the vehicle 102. For example, the controller 116 can issue commands to the drive controller 118 for the vehicle 102 to move according to a guidance line generated by the controller 116. In such examples, the drive controller 118 responds to these commands and causes movement of the vehicle 102 (e.g., movement of the wheels, steering adjustments, etc.).
The example implement controller 120 of the illustrated example of FIG. 1 causes the implement 108 to perform an operation on the field. For example, the implement controller 120 can cause the implement 108 to plant, spray, harvest, and/or perform any other operation. In some examples, the implement controller 120 causes portions of the implement 108 to perform an operation, based on data from the controller 116 indicating that some of the rows under the implement 108 have already been operated upon (e.g., as determined based on mapping data). In some examples, the drive controller 118 and/or the implement controller 120 may be integrated into the controller 116.
FIG. 2 is a schematic of the controller 116 of the vehicle 102 of FIG. 1. The controller 116 includes an example field map handler 202, an example location data analyzer 204, an example field map generator 206, an example map data store 208, an example field map analyzer 210, an example user command accessor 212, an example user interface configurator 214, an example equipment parameter determiner 216, an example guidance line generator 218, an example implement commander 220, an example drive commander 222, and an example operation performance tracker 224.
The example field map handler 202 of the illustrated example of FIG. 2 analyzes location data and generates and stores field maps based on the location data. The field map handler 202 receives GPS data and/or inertial data from the location sensor 114. If a field map already exists for a location indicated in the location data (e.g., indicating the current operation is a subsequent operation), then the field map handler 202 can pass the location data (e.g., the GPS data and/or the inertial data) to the field map analyzer 210, which can retrieve the field map from the map data store 208. Conversely, if a map has not been generated, the field map generator 206 can generate a new field map based on the location data. The field map handler 202 of the illustrated example includes an example location data analyzer 204, an example field map generator 206, and an example map data store 208.
The example location data analyzer 204 of the illustrated example of FIG. 2 analyzes location data from the location sensor 114. In some examples, the location data analyzer 204 receives GPS data and inertial data from the location sensor 114, and analyzes the GPS data and inertial data to determine precise locations of the vehicle 102 and/or the implement 108. In some examples, the location data from the location sensor 114 already directly indicates locations of the vehicle 102 and/or the implement 108. The location data analyzer 204 of the illustrated example determines, based on the location data, locations of portions of the implement 108 corresponding to individual ones of the crop rows 110. In some examples, the location data analyzer 204 queries the map data store 208 to determine whether a field map exists for the current location of the vehicle 102. In some examples, the location data analyzer 204 determines whether a field map exists within a specified time period (e.g., a field map from the past six months, a field map from the past year, etc.), as outdated maps (outside the time period) may no longer be accurate if a field has been re-planted. In some examples, the location data analyzer 204 communicates location data to the field map analyzer 210 if a field map corresponding to the location data already exists.
The example field map generator 206 of the illustrated example of FIG. 2 generates field maps based on location data from the location sensor 114. In some examples, the location data analyzer 204 analyzes the location data to determine locations of the vehicle 102 and/or portions of the implement 108, which are then received by the field map generator 206. The field map generator 206 of the illustrated example creates a new map when location data is received for a location for which no map data, or no recent map data (e.g., from the current agricultural season) exists. The field map generator 206 of the illustrated example stores (a) location data including the guidance line which the vehicle 102 followed while performing the initial operation (e.g., planting, tilling, etc.). (b) locations of portions of the implement 108 during the operation, and/or (c) as data from the implement sensors 112 a-l indicating whether an operation was performed by portions of the implement 108.
In some examples, the operation performance tracker 224 updates the field map generated by the field map generator 206 based on data from the implement sensors 112 a-l. For example, the operation performance tracker 224 may indicate that three portions of the implement 108 were not utilized during a pass through a field. In some such examples, the field map generator 206 updates the field map to indicate the operation did not occur on the crop rows corresponding to these three portions of the implement 108 during the specified pass through the field. Therefore, the field map generated by the field map generator 206 can indicate (a) precise locations (e.g., accounting for displacements in all three-axes) of the vehicle's guidance line that was followed during an initial operation, (b) precise locations of portions of the implement 108 corresponding to crop rows, and/or (c) indications of whether portions of the implement 108 were utilized during the initial operation.
The field map generator 206 stores the field map in the map data store 208 and/or another location accessible to the controller 116. In some examples, field maps may be stored at a location accessible via a network, thereby enabling multiple vehicles to access field maps (e.g., in case multiple vehicles perform a coordinated operation, in case different vehicles are utilized in subsequent operations, etc.).
In some examples where an operation is executed by multiple vehicles in coordination, the field map generator 206 may combine mapping data (e.g., location data, implement operation data, etc.) from multiple vehicles into a common map for a field. In some such examples, the field map generator 206 may be implemented on a network and/or at a central computing system. In some examples, the field map generator 206 of one vehicle, such as the vehicle 102, serves as a primary field map generator when multiple vehicles are performing a coordinated operation. In some examples, individual vehicles performing a coordinated operation may calculate guidance lines and perform other guidance calculations independently, but based upon the same set of rules, such that the vehicles operate predictably and will be coordinated in performing the operation.
The map data store 208 of the illustrated example of FIG. 2 stores field maps, which can be subsequently accessed to enable automated operation of the vehicle 102 (e.g., via generation and utilization of accurate, crop row-specific guidance lines). In some examples, the map data store 208 is accessible via a network. The map data store 208 may be implemented by a volatile memory (e.g., a Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM), etc.) and/or a nonvolatile memory (e.g., flash memory). The map data store 208 may additionally or alternatively be implemented by one or more double data rate (DDR) memories, such as DDR, DDR2, DDR3, mobile DDR (mDDR), etc. The map data store 208 may additionally or alternatively be implemented by one or more mass storage devices such as hard disk drive(s), compact disk drive(s) digital versatile disk drive(s), etc. While in the illustrated example the map data store 208 is illustrated as a single database, the map data store 208 may be implemented by any number and/or type(s) of databases. Furthermore, the data stored in the map data store 208 may be in any data format such as, for example, binary data, comma delimited data, tab delimited data, structured query language (SQL) structures, etc. The location data analyzer 204, the field map generator 206, the field map analyzer 210, and/or the operation performance tracker 224 may access and/or modify field maps stored in the map data store 208.
The example field map analyzer 210 of the illustrated example of FIG. 2 analyzes field maps stored in the map data store 208. The field map analyzer 210 can retrieve field maps based on location data from the location data analyzer 204. The field map analyzer 210 can retrieve field maps for any location in a field, thereby enabling the guidance line generator 218 to generate guidance lines at any location. Therefore, the field map analyzer 210 enables the vehicle 102 to begin an operation at any location in a field and accurately operate on the field based on precise locations of the crop rows 110 a-1 represented in the field map. The field map analyzer 210 of the illustrated example determines crop row locations based on field maps retrieved from the map data store 208. In some examples where multiple vehicles perform a coordinated operation, the field map analyzer 210 determines appropriate portions of a field map to communicate with individual vehicles executing operations on portions of the field.
The example user command accessor 212 of the illustrated example of FIG. 2 accesses commands issued by an operator. For example, the user command accessor 212 can receive commands entered by an operator via the user interface 122. The user command accessor 212 can receive commands to perform various tasks, such as initiating or terminating an automated operation on a field, manually adjusting guidance lines for the vehicle 102 to travel along, initiating or terminating collection of data (e.g., location data, operation performance data, etc.), manually adjusting a setting of the implement (e.g., a spray power, a harvest speed, etc.), and/or any other tasks associated with operation of the vehicle 102. The user command accessor 212 analyzes command(s) entered by an operator and causes implementation of the command(s). For example, if an operator enters a command to cease all data collection, the user command accessor 212 can communicate with the operation performance tracker 224 to cease analysis and storage of implement operation data and the location data analyzer 204 to cease analysis and storage of location data (e.g., GPS data, inertial data, etc.). Example commands which may be received at the user command accessor 212 include those that may be issued from the run screen 702 illustrated and described in association with FIG. 7.
The example user interface configurator 214 of the illustrated example of FIG. 2 configures the user interface 122. For example, the user interface configurator 214 can communicate data and/or images to be displayed on the user interface 122. In some examples, the user interface configurator 214 communicates map data from the field map analyzer 210 and/or the map data store 208 to the user interface configurator 214 to cause the user interface 122 to display a field map, crop row locations, and/or other data included in a field map. In some examples, the user interface configurator 214 accesses guidance lines generated by the guidance line generator 218 and causes the user interface 122 to display the guidance lines over the field map for a specific operation (e.g., planting, cultivating, spraying, etc.). In some examples, the user interface configurator 214 communicates operation performance data from the operation performance tracker 224 to be displayed on the user interface 122. In some such examples, the user interface 122 may display color and/or shading indicating areas in which an operation has been performed. An example user interface configured by the user interface configurator 214 is illustrated and described in association with FIG. 7.
The example equipment parameter determiner 216 of the illustrated example of FIG. 2 determines parameters of the implement 108. For example, the equipment parameter determiner 216 may determine a type of the implement 108 (e.g., harvester, sprayer, planter, etc.), an operational width of the implement 108 (e.g., one-row, four-rows, eight-rows, etc.), and/or any other parameters for the implement 108. In some examples, the equipment parameter determiner 216 automatically receives data from the implement 108 when the implement 108 is connected to the vehicle 102. In some examples, the user command accessor 212 accesses commands including parameters association with the implement 108. For example, an operator may input a type and/or width of the implement into the user interface 122, which can then be accessed by the user command accessor 212 and communicated to the equipment parameter determiner 216.
In some examples, the equipment parameter determiner 216 communicates with the guidance line generator 218 to enable generation of guidance lines that are specific to parameters of the implement 108. For example, if a field to be operated upon includes thirty-six rows according to the field map as analyzed by the field map analyzer 210, and the equipment parameter determiner 216 determines that the implement 108 can operate on four rows at a time, the guidance line generator 218 may cause nine guidance lines to be generated to ensure all rows are operated upon.
The example guidance line generator 218 of the illustrated example of FIG. 2 generates guidance lines that are specific to the implement 108 attached to the vehicle 102 and/or specific to the field map corresponding to a current field being operated upon. The guidance line generator 218 of the illustrated example generates guidance lines to ensure the vehicle 102 moves along a path that allows the implement 108 to perform its designated operation (e.g., planting, spraying, fertilizing, etc.) most efficiently. For example, the guidance line generator 218 generates as few guidance lines as possible to cover an entire field. In some examples, the guidance line generator 218 generates guidance lines to avoid wheels of the vehicle 102 or wheels of the implement 108 running over crops rows. For example, sugar cane crops are susceptible sustaining significant permanent damage (as it is a perennial crop) if run over. Hence, the guidance line generator 218 of the illustrated example generates guidance lines that ensure accurate, efficient and safe guidance lines for the vehicle 102 to follow when operating autonomously. The equipment parameter determiner 216 receives information pertaining to crop rows identified in field maps by the field map analyzer 210 and information pertaining to equipment parameters from the equipment parameter determiner 216. The guidance line generator 218 outputs guidance lines to the drive commander 222 and the implement commander 220 to enable the drive commander 222 to follow the guidance lines when the vehicle 102 is operating in an autonomous mode, and to enable the implement commander 220 to foresee the route of the vehicle 102 and cause the implement 108 to perform efficient operations (e.g., not repeating operations on rows that are passed over twice). The guidance line generator 218 of the illustrated example communicates the guidance lines to the user interface configurator 214 to enable an operator to view the guidance lines on the user interface 122.
The implement commander 220 of the illustrated example of FIG. 2 causes the implement 108 to perform operations on the field. The implement commander 220 actuates portions of the implement 108 to cause an operation (e.g., planting, tilling, spraying) to be executed by the portions of the implement 108. The implement commander 220 utilizes field map data from the field map analyzer 210 and guidance line data from the guidance line generator 218 to cause portions of the implement 108 to perform operations. For example, if the implement commander 220 receives guidance line data indicating that a centerline of the vehicle 102 will be traveling between rows five and six, the implement commander 220 can reference corresponding field map data from the field map analyzer to determine whether the current operation (e.g., corresponding to the current type of implement) has already been performed on rows that will be passed by the implement 108. In some such examples, the implement 108 can selectively activate portions of the implement 108 to intelligently perform operations on only those rows which have not yet been operated upon. The implement commander 220 of the illustrated example communicates implement control commands to the implement controller 120. In some examples, the implement commander 220 can send implement control commands to the implement controller 120 based on manual operator inputs (e.g., implement controls manually actuated by the operator).
The drive commander 222 of the illustrated example of FIG. 2 issues drive control commands to the drive controller 118 to cause the vehicle 102 to move. In some examples, the drive commander 222 issues the drive control commands to cause the vehicle 102 to move along a guidance line generated by the guidance line generator 218. In some examples, the drive commander 222 issues drive control commands based on user commands input into the user interface 122 (e.g., as accessed by the user command accessor 212) or commands otherwise issued form another control in the cab 106. The drive commander 222 of the illustrated example causes components of the vehicle 102 to adjust to terrain represented in the field map analyzed by the field map analyzer 210 while traversing a guidance line. In some examples, the drive commander 222 may operate autonomously (e.g., cause the vehicle 102 to automatically follow guidance lines generated by the guidance line generator 218), while the implement commander 220 may be manually operated (e.g., may be responsive to manual operator inputs), or vice-versa. In some examples, in a fully automated mode, both the drive commander 222 and the implement commander 220 operate autonomously based on location data, field map data, equipment parameters, guidance lines, and/or implement operation data.
The operation performance tracker 224 of the illustrated example of FIG. 2 analyzes implement operation data from the implement sensors 112 a-l. The operation performance tracker 224 of the illustrated example integrates operation performance data into one or more field maps corresponding to a field currently under operation. For example, if the implement operation data indicates that a portion of the crop rows 110 a-c was operated upon, the data from the corresponding implement sensors 112 a-c can be analyzed by the operation performance tracker 224, and the field map corresponding to a current location (e.g., as determined by the location data analyzer) can be updated to include data indicating these rows have upon operated upon. In such examples, the implement commander 220 can intelligently operate the implement 108 to avoid performing a repeat operation on these rows 110 a-c. The operation performance tracker 224 may directly modify field maps from the map data store 208 and/or another storage location, or may communicate implement operation data to the field map analyzer 210, the field map generator 206, and/or the implement commander 220.
In operation, the field map handler 202 enables generation of field maps by analyzing location data using the location data analyzer 204, generating maps based on the location data using the field map generator 206, and storing the field maps for subsequent use in the map data store 208. When a vehicle 102 travels over a field that has been mapped, the field map analyzer 210 can analyze the existing map and determine crop row locations and other data to enable automated navigation and operation throughout the field. The user command accessor 212 can process a variety of commands from an operator via the user interface 122 or another component in the cab 106. The user interface configurator 214 communicates user interface data to the user interface 122 such that field maps and other data can be displayed on the user interface 122 for the operator. The equipment parameter determiner 216 determines parameters (e.g., a type of implement, an operational width of an implement, etc.) of the implement 108. The guidance line generator 218 utilizes field map(s) and/or equipment parameters to generate guidance lines for the vehicle 102, which are subsequently employed by the implement commander 220 and the drive commander 222 to enable automated navigation and operation of the vehicle 102. The operation performance tracker 224 monitors implement operation data to update field maps based on operations that have been performed.
While an example manner of implementing the controller 116 of FIG. 1 is illustrated in FIG. 2, one or more of the elements, processes and/or devices illustrated in FIG. 2 may be combined, divided, re-arranged, omitted, eliminated and/or implemented in any other way. Further, the example field map handler 202, the example location data analyzer 204, the example field map generator 206, the example map data store 208, the example field map analyzer 210, the example user command accessor 212, the example user interface configurator 214, the example equipment parameter determiner 216, the example guidance line generator 218, the example implement commander 220, the example drive commander 222, the example operation performance tracker 224 and/or, more generally, the example controller 116 of FIG. 2 may be implemented by hardware, software, firmware and/or any combination of hardware, software and/or firmware. Thus, for example, any of the example field map handler 202, the example location data analyzer 204, the example field map generator 206, the example map data store 208, the example field map analyzer 210, the example user command accessor 212, the example user interface configurator 214, the example equipment parameter determiner 216, the example guidance line generator 218, the example implement commander 220, the example drive commander 222, the example operation performance tracker 224 and/or, more generally, the example controller 116 of FIG. 2 could be implemented by one or more analog or digital circuit(s), logic circuits, programmable processor(s), programmable controller(s), graphics processing unit(s) (GPU(s)), digital signal processor(s) (DSP(s)), application specific integrated circuit(s) (ASIC(s)), programmable logic device(s) (PLD(s)) and/or field programmable logic device(s) (FPLD(s)). When reading any of the apparatus or system claims of this patent to cover a purely software and/or firmware implementation, at least one of the example field map handler 202, the example location data analyzer 204, the example field map generator 206, the example map data store 208, the example field map analyzer 210, the example user command accessor 212, the example user interface configurator 214, the example equipment parameter determiner 216, the example guidance line generator 218, the example implement commander 220, the example drive commander 222, the example operation performance tracker 224 and/or, more generally, the example controller 116 of FIG. 2 is/are hereby expressly defined to include a non-transitory computer readable storage device or storage disk such as a memory, a digital versatile disk (DVD), a compact disk (CD), a Blu-ray disk, etc. including the software and/or firmware. Further still, the example controller 116 of FIG. 3 may include one or more elements, processes and/or devices in addition to, or instead of, those illustrated in FIG. 2, and/or may include more than one of any or all of the illustrated elements, processes and devices. As used herein, the phrase “in communication,” including variations thereof, encompasses direct communication and/or indirect communication through one or more intermediary components, and does not require direct physical (e.g., wired) communication and/or constant communication, but rather additionally includes selective communication at periodic intervals, scheduled intervals, aperiodic intervals, and/or one-time events.
FIG. 3 is a schematic illustrating an example coverage-based mapping technique and an example row-based mapping technique. The schematic includes an example coverage map 302 and an example row map 304. The example coverage map 302 includes shading on areas where an operation has been performed. In the example coverage map 302, rectangular areas are utilized as approximations of where the implement 108 performed an operation. During subsequent use of the coverage map 302 to generate guidance lines for autonomous operation of the vehicle 102, the guidance line generator 218 would not be able to identify locations of specific crop rows, as the coverage map 302 merely indicates generally where an operation was performed.
The example row map 304 of the illustrated example of FIG. 3 includes example crop row lines 308 a-d, which illustrate precise locations of crop rows on the field. The row map 304 further includes an example guidance line 310, illustrating a path that the vehicle 102 followed or is to follow while traversing the field. The crop row lines 308 a-d can be integrated in the field map by utilizing data from the operation performance tracker 224, the equipment parameter determiner 216, and/or the location data analyzer 204. For example, the location data analyzer 204 can determine a location of the vehicle 102 based on location data from the location sensor 114. Accurate positions of individual portions of the implement 108 can be determined by extrapolating the location data for the vehicle 102 based on a known width of the implement and a known location of the location sensor 114 relative to the implement 108. Finally, the operation performance tracker 224 can determine whether portions of the implement 108 were actually utilized during an operation, which therefore determines whether those crop rows should exist in a new field map. For example, if a map is generated during an initial planting operation, the crop row lines 308 a-d should only be generated at locations of crop rows that are actually planted, according to the operation performance tracker 224.
Utilizing a row-based mapping technique allows subsequent generation of accurate guidance lines to perform an operation upon the crop rows using a different implement (which may have a different width) at a later time. Conversely, utilizing a coverage-based map which only depicts general areas which have been operated upon may result in guidance lines being generated that cause overlap of operations, damage to crops, and/or crop rows remaining untreated during subsequent operations.
FIG. 4A is an example field map 402 generated using an example four-row implement. The field map 402 of the illustrated example of FIG. 4A includes a plurality of example crop row lines 404 and a plurality of example original guidance lines 406. The crop row lines 404 accurately illustrate locations of crop rows on the field. For example, the crop row lines 404 can accurately represent crop row locations based on location data determined directly on the implement, therefore accounting for potential implement drift during the operation. The original guidance lines 4 depict a path that was followed by the vehicle which generated the field map. For each one of the original guidance lines 406, there are four crop row lines 404 distributed around the one of the original guidance lines 406, indicating that the implement of the original vehicle that performed the mapping had an operational width of four-rows. While only one of the crop row lines 404 and one of the original guidance lines 406 are labeled, all lines in FIGS. 4A-4F which have the same line type (using the same dashing pattern) are understood to refer to the same type of element.
FIG. 4B is an example first field map traversal schematic 408 using an example twelve-row sprayer 410 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A. The first field map traversal schematic 408 includes an example guidance line 412 for the twelve-row sprayer 410. The twelve-row sprayer 410 has an operational width that is three times larger than the operational width of the implement which performed the original operation. Thus, there is only the guidance line 412 in the first field map traversal schematic 408 in order for the twelve-row sprayer to perform a spraying operation on twelve of the crop row lines 404.
FIG. 4C is an example second field map traversal schematic 414 using an example eight-row sprayer 416 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A. The second field map traversal schematic 414 includes an example guidance line 418 for the eight-row sprayer 416. The eight-row sprayer 416 has an operational width that is two times larger than the operational width of the implement which performed the original operation. Thus, there is only the guidance line 418 in the second field map traversal schematic 414 in order for the eight-row sprayer 416 to perform a spraying operation on eight of the crop row lines 404.
FIG. 4D is an example third field map traversal schematic 420 using an example four-row harvester 422 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A. The third field map traversal schematic 420 includes example guidance lines 424 for the four-row harvester 422. While the guidance lines 424 of the third field map traversal schematic 420 substantially overlap the original guidance lines 406, in some examples, due to implement drift or other factors, the actual crop row locations may be different such that the guidance lines 424 are different from the original guidance lines 406 to best operate upon the crop rows, despite the operational width of the four-row harvester 422 being the same as the four-row implement of the original operation.
FIG. 4E is an example fourth field map traversal schematic 426 using an example two-row harvester 428 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A. The fourth field map traversal schematic 426 includes example guidance lines 430 for the two-row harvester 428. Since the two-row harvester 428 has an operational width that is half that of the four-row implement used in the original operation, there are twice as many of the guidance lines 430 in the fourth field map traversal schematic 426 relative to the original guidance lines 406.
FIG. 4F is an example fifth field map traversal schematic 432 using an example one-row harvester 434 in a subsequent operation on the field corresponding to the field map 402 of FIG. 4A. The fifth field map traversal schematic 432 includes example guidance lines 436 which are displayed directly over the crop row lines 404 of the field map 402, since the one-row harvester 434 must individually pass over each crop row to harvest the field. The one-row harvester 434 therefore requires four times as many guidance lines as the four-row implement used in the original operation.
FIG. 5A is an enlarged view of the example field map 402 of FIG. 4A generated using a four-row implement.
FIG. 5B is an example sixth field map traversal schematic 502 using an example single row harvester 504 with an example haulage vehicle 506 on a portion of the field corresponding to the field map 402 of FIG. 5A. The single row harvester 504 is a first implement which harvests the crop rows, and the haulage vehicle 506 is a second implement which is used to store the harvest during operation. Hence, the controller 116 must generate guidance lines for the vehicle based on operational widths and relative positioning of both the single row harvester 504 and the haulage vehicle 506. For clarity, the guidance lines for the single row harvester 504 and the haulage vehicle 506 are omitted from FIG. 5B, and instead illustrated with each of the implements individually in FIGS. 5C and 5D. While individual guidance lines are illustrated for each of the implements (the single row harvester 504 and the haulage vehicle 506) in FIGS. 5C, 5D, the controller (e.g., the guidance line generator 218) may generate a single guidance line for the overall vehicle to follow based on the individual guidance lines for the implements.
FIG. 5C is an example seventh field map traversal schematic 508 using the example single row harvester 504 of FIG. 5B on the field corresponding to the field map 402 of FIG. 5A. Since the single row harvester 504 is only capable of operating on one row at a time, example guidance lines 510 of FIG. 5C for the single row harvester 504 overlap on all crop rows of the field map 402 to ensure all crop rows are operated upon. While the guidance lines 510 of the illustrated example substantially overlap the original guidance lines 406, in some examples, even when the operational width of the vehicle implement performing the subsequent operation is the same as the operational width of the vehicle implement performing the original operation, the guidance lines may not overlap. For example, due to drift of the implement, crop row locations may not be evenly distributed around the original guidance line. In some such examples, when a subsequent operation is performed, the guidance lines may be generated in different locations than the original guidance lines to better target the actual locations of the crop rows. Thus, the benefit of having knowledge of specific locations of the crop rows can be realized even in examples where a subsequent vehicle has a same operational width as an original vehicle.
FIG. 5D is an example eighth field map traversal schematic 512 using the example haulage vehicle 506 of FIG. 5B on the field corresponding to the field map 402 of FIG. 5A. The eighth field map traversal schematic 512 includes example guidance lines 514 for the haulage vehicle 506. The guidance lines 514 of the illustrated example of FIG. 5D are generated (e.g., by the guidance line generator 218) such that wheels of the haulage vehicle 506 do not drive over the crop rows represented by the crop row lines 404. The haulage vehicle 506 does not perform an operation on the crop rows itself, but instead assists the single-row harvester 504. Hence, the guidance lines 514 of FIG. 5B are not generated to enable an operation on the crop rows, but to avoid damaging the crop rows (e.g., by driving over them with the haulage vehicle 506).
FIG. 6A is an example ninth field map traversal schematic 602 illustrating an example original operation and an example subsequent operation executed on a field map with a narrow guess row. The ninth field map traversal schematic 602 includes an original vehicle 604 which performed the original operation and generate the field map. The two instances of the original vehicle 604 represents subsequent passes of the same original vehicle 604. The ninth field map traversal schematic 602 includes example crop row lines 606 indicating locations at which the crop rows were operated upon (e.g., planted) during the original operation. In the ninth field map traversal schematic 602, the original operation exhibited an example narrow guess row 608, meaning that outer crop rows from consecutive passes of the original vehicle 604 have a narrow spacing relative to other crop rows. Without accurate knowledge of the crop row locations (e.g., as provided by field maps generated using techniques disclosed herein), subsequent operations performed on the field may drive over the narrow guess row 608, and/or otherwise improperly operate upon the crop rows of the narrow guess row 608.
Utilizing field mapping techniques disclosed herein, a subsequent operation can accurately generate guidance lines to appropriately operate upon a field with a narrow guess row. For example, the ninth field map traversal schematic 602 includes an example subsequent vehicle 610 with a twelve-row implement. The twelve-row implement could theoretically operate upon all of the crop rows represented by the crop row lines 606, but to do so, the subsequent vehicle 610 would have to drive in the center of the crop rows, and may end up driving over the narrow guess row 608. To avoid this, example guidance lines 612 are generated accounting for the specific positions of the crop rows and the width of the implement of the subsequent vehicle 610. The example guidance lines 612 cause the subsequent vehicle 610 to travel between rows with regular widths, thus avoiding driving through the narrow guess row 608. In such examples, the implement commander 220 of the controller 116 can cause portions of the implement to be enabled and/or disabled based on whether crop rows beneath the implement have been operated upon. Based on the positioning of the guidance lines 612, utilizing all portions of the implement throughout the entire traversal would result in repeated operations, and hence, the implement must be selectively activated over rows which have not yet been operated upon.
FIG. 6B is an example tenth field map traversal schematic 614 illustrating another original operation and another subsequent operation executed on a field map with an accurate guess row. The tenth field map traversal schematic 614 includes the example original vehicle 604 and the example subsequent vehicle 610. In the tenth field map traversal schematic 614, the original operation is performed by the original vehicle 604 with example crop row lines 606 including an example accurate guess row 618 (e.g., a guess row where the spacing between outer crop rows operated upon by the original implement during consecutive passes are spaced approximately equally to other crop rows). In the illustrated example of FIG. 6B, the subsequent vehicle 610 utilizes an example guidance line 620 that is in between crop rows of the accurate guess row 618. The guidance line generator 218 of the illustrated example of FIG. 2 can generate the guidance line 620 centrally such that the twelve-row implement of the subsequent vehicle 610 can operate upon all of the twelve crops rows without the subsequent vehicle 610 running over any of the crop rows (e.g., as may happen utilizing the guidance line 620 of FIG. 6B with the narrow guess row of FIG. 6A. Hence, guidance lines generated for automated vehicle operation utilizing techniques disclosed herein enable accurate and efficient operation on afield regardless of unique spacing of crop rows (e.g., regardless of differently-sized guess rows, regardless of contours of the field, etc.).
FIG. 7 is an example run screen 702 to be displayed on the example user interface 122 of the vehicle 102 of FIG. 1. The example run screen 702 includes an example navigation display portion 704 to display mapping data, guidance lines, crop row locations, and/or other information pertinent to navigation of the vehicle 102.
The navigation display portion 704 includes an example vehicle position indicator 706 displaying a current position of the vehicle 102 relative to the field illustrated in the navigation display portion 704.
The navigation display portion 704 further includes example crop row lines 708, illustrating precise locations of crop rows, as determined from a previously generated field map (e.g., generated by the field map generator 206 during an initial operation).
The navigation display portion 704 includes an example current guidance line 710 (displayed as a solid line) that the vehicle 102 is following during its automated operation on the field. In some examples, the current guidance line 710 is generated by the guidance line generator 218 of the controller 116 of FIG. 2. Similarly, the navigation display portion 704 includes example past or future guidance lines 712, which indicate guidance lines that are either not yet being followed or were already followed during previous operation.
The navigation display portion 704 includes an example implement position line 714, indicating a current position and operational width of the implement 108. Behind the implement (e.g., under the implement, in the view of FIG. 7) on the field is an example coverage section 716. The coverage section 716 provides shading for areas which have been operated upon by the implement 108. In some examples, the operation performance tracker 224 indicates portions of the implement 108 which have been enabled throughout an operation. In some such examples, this implement operation data can be combined with location data indicating a position of the implement 108 to plot the coverage section 716.
The run screen 702 of the illustrated example of FIG. 7 includes an example guidance portion 718. The guidance portion 718 includes an example set track button 720 to allow an operate to determine a track (e.g., one or more guidance lines) for the vehicle 102 to follow. In some examples, the tracks (e.g., the plurality of guidance lines) are determined automatically based on a location of the vehicle 102 and existing field maps. In some examples, the set track button 720 allows the user to select a guidance line and/or begin the process of manually creating a new guidance line (e.g., in the absence of mapping data). The guidance portion 718 further includes an example track adjustment portion 722 to enable manual adjustments of the guidance line. For example, if the current guidance line 710 appears to be problematic (e.g., the operator notices that the vehicle 102 is driving over a crop row), the operator can manually shift a position of the current guidance line 710.
The run screen 702 of the illustrated example of FIG. 7 further includes an example setup button 724. The setup button 724, when pressed, allows an operator to adjust parameters of the vehicle 102 and/or the implement 108. For example, if the implement 108 does not automatically communicate parameters (e.g., an operational width, an implement type, etc.) to the controller 116, the operator can enter these parameters into a prompt provided when the setup button 724 is pressed.
The run screen 702 of the illustrated example of FIG. 7 includes an example work recording button 726. When actuated to “on,” as shown in the illustrated example of FIG. 7, the controller 116 records implement operation data, thereby enabling generation of the coverage section 716. In some examples, the operation performance tracker 224 communicates operation performance data to the map data store 208, the field map analyzer 210, and/or the field map generator 206, such that the coverage section 716 can be stored in the field map and accessed by an operator to view portions of the field which have been operated upon.
The run screen 702 of the illustrated example of FIG. 7 includes an example auto-trac button 728, which allows an operator to enable or disable automated guidance of the vehicle 102 and/or automated operation of the implement 108. Since the auto-trac button is listed as enabled in FIG. 7, the vehicle 102 is operating in an automated mode, automatically generating and following guidance lines and performing operations with the implement 108 based on existing field maps.
Flowcharts representative of example hardware logic, machine readable instructions, hardware implemented state machines, and/or any combination thereof for implementing the controller 116 of FIGS. 1 and 2 are shown in FIGS. 8-10. The machine readable instructions may be one or more executable programs or portion(s) of an executable program for execution by a computer processor such as the processor 1112 shown in the example processor platform 1100 discussed below in connection with FIG. 11. The program may be embodied in software stored on a non-transitory computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a DVD, a Blu-ray disk, or a memory associated with the processor 1112, but the entire program and/or parts thereof could alternatively be executed by a device other than the processor 1112 and/or embodied in firmware or dedicated hardware. Further, although the example program is described with reference to the flowcharts illustrated in FIGS. 8-10, many other methods of implementing the example controller 116 may alternatively be used. For example, the order of execution of the blocks may be changed, and/or some of the blocks described may be changed, eliminated, or combined. Additionally or alternatively, any or all of the blocks may be implemented by one or more hardware circuits (e.g., discrete and/or integrated analog and/or digital circuitry, an FPGA, an ASIC, a comparator, an operational-amplifier (op-amp), a logic circuit, etc.) structured to perform the corresponding operation without executing software or firmware.
The machine readable instructions described herein may be stored in one or more of a compressed format, an encrypted format, a fragmented format, a packaged format, etc. Machine readable instructions as described herein may be stored as data (e.g., portions of instructions, code, representations of code, etc.) that may be utilized to create, manufacture, and/or produce machine executable instructions. For example, the machine readable instructions may be fragmented and stored on one or more storage devices and/or computing devices (e.g., servers). The machine readable instructions may require one or more of installation, modification, adaptation, updating, combining, supplementing, configuring, decryption, decompression, unpacking, distribution, reassignment, etc. in order to make them directly readable and/or executable by a computing device and/or other machine. For example, the machine readable instructions may be stored in multiple parts, which are individually compressed, encrypted, and stored on separate computing devices, wherein the parts when decrypted, decompressed, and combined form a set of executable instructions that implement a program such as that described herein. In another example, the machine readable instructions may be stored in a state in which they may be read by a computer, but require addition of a library (e.g., a dynamic link library (DLL)), a software development kit (SDK), an application programming interface (API), etc. in order to execute the instructions on a particular computing device or other device. In another example, the machine readable instructions may need to be configured (e.g., settings stored, data input, network addresses recorded, etc.) before the machine readable instructions and/or the corresponding program(s) can be executed in whole or in part. Thus, the disclosed machine readable instructions and/or corresponding program(s) are intended to encompass such machine readable instructions and/or program(s) regardless of the particular format or state of the machine readable instructions and/or program(s) when stored or otherwise at rest or in transit.
As mentioned above, the example processes of FIGS. 8-10 may be implemented using executable instructions (e.g., computer and/or machine readable instructions) stored on a non-transitory computer and/or machine readable medium such as a hard disk drive, a flash memory, a read-only memory, a compact disk, a digital versatile disk, a cache, a random-access memory and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information). As used herein, the term non-transitory computer readable medium is expressly defined to include any type of computer readable storage device and/or storage disk and to exclude propagating signals and to exclude transmission media.
“Including” and “comprising” (and all forms and tenses thereof) are used herein to be open ended terms. Thus, whenever a claim employs any form of “include” or “comprise” (e.g., comprises, includes, comprising, including, having, etc.) as a preamble or within a claim recitation of any kind, it is to be understood that additional elements, terms, etc. may be present without falling outside the scope of the corresponding claim or recitation. As used herein, when the phrase “at least” is used as the transition term in, for example, a preamble of a claim, it is open-ended in the same manner as the term “comprising” and “including” are open ended. The term “and/or” when used, for example, in a form such as A, B, and/or C refers to any combination or subset of A, B, C such as (1) A alone, (2) B alone, (3) C alone, (4) A with B. (5) A with C. (6) B with C, and (7) A with B and with C. As used herein in the context of describing structures, components, items, objects and/or things, the phrase “at least one of A and B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B. and (3) at least one A and at least one B. Similarly, as used herein in the context of describing structures, components, items, objects and/or things, the phrase “at least one of A or B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B, and (3) at least one A and at least one B. As used herein in the context of describing the performance or execution of processes, instructions, actions, activities and/or steps, the phrase “at least one of A and B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B, and (3) at least one A and at least one B. Similarly, as used herein in the context of describing the performance or execution of processes, instructions, actions, activities and/or steps, the phrase “at least one of A or B” is intended to refer to implementations including any of (1) at least one A, (2) at least one B and (3) at least one A and at least one B.
Example machine readable instructions 800 that may be executed by the controller 116 to perform an operation on a field using a field map are illustrated in FIG. 8. With reference to the preceding figures and associated description, the example machine readable instructions 800 begin with the example controller 116 determining equipment parameter(s) automatically or via manual entry (Block 802). In some examples, the equipment parameter determiner 216 determines equipment parameter(s) automatically (e.g., based on data from the implement 108 and/or from the implement sensors 112 a-1), and/or via manual entry (e.g., via entry from the operator on the user interface 122).
At block 804, the example controller 116 determines if the current operation is the first operation on the field. In some examples, the field map analyzer 210 determines if the current operation is the first operation on the field by comparing a location of the vehicle 102, as determined by the location data analyzer 204, with existing field maps in the map data store 208. If a field map exists, then the current operation is not the first operation on the field. Conversely, if a field map does not exist, the current operation is the first operation on the field. In response to the current operation being the first operation on the field, processing transfers to block 806. Conversely, in response to the current operation not being the first operation on the field, processing transfers to block 808.
At block 806, the example controller 116 generates a field map during a first operation. Further detail of the implementation of block 806 is illustrated and described below in connection with FIG. 9.
At block 808, the example controller 116 retrieves a map associated with the field. In some examples, the field map analyzer 210 retrieves a field map from the map data store 208 and/or form another location (e.g., a network location) accessible to the controller 116.
At block 810, the example controller 116 determines crop row locations. In some examples, the field map analyzer 210 determines crop row locations based on the field map. In some examples, the field map analyzer 210 determines crop row locations based on the field map and historical implement operation data, as analyzed by the operation performance tracker 224.
At block 812, the example controller 116 determines guidance line(s) based on crop row locations and implement parameters. In some examples, the guidance line generator 218 generates guidance lines based on crop row locations from the field map analyzer 210 and implement parameters from the equipment parameter determiner 216. For example, the guidance line generator 218 can utilize a known operational width of the implement 108 to generate guidance lines to perform an operation upon the crop rows.
At block 814, the example controller 116 displays mapping data and a guidance line on the user interface 122. In some examples, the user interface configurator causes the user interface 122 to display mapping data (e.g., a field map) and a guidance line on the user interface 122. An example of such mapping data and a guidance line displayed on the user interface 122 is illustrated by the run screen 702 of FIG. 7.
At block 816, the example controller 116 automatically guides the vehicle 102 along guidance line(s) and perform operation(s). Further detail of the implementation of block 816 is illustrated and described below in connection with FIG. 10.
At block 818, the example controller 116 determines whether the operation is complete. In some examples, the user command accessor 212 determines, based on user commands entered on the user interface 122, whether the operation is complete. In some examples, the operation performance tracker 224 determines, based on implementation operation data and a location of the vehicle 102, whether the operation is complete. In response to the operation being complete, processing transfers to block 820. Conversely, in response to the operation not being complete, processing transfers to block 814.
At block 820, the example controller 116 determines whether there is an additional operation to perform. In some examples, the user command accessor 212 determines, based on user commands, whether there is an additional operation to perform on the field. In response to there being an additional operation to perform, processing transfers to block 802. Conversely, in response to there not being an additional operation to perform, processing transfers to block 822.
At block 822, the example controller 116 saves implement operation data in association with the field map. In some examples, the operation performance tracker 242, the field map analyzer 210, the field map generator 206, and/or the map data store 208 save implement operation data in association with the field map.
Example machine readable instructions 900 that may be executed by the controller 116 to generate a first operation map are illustrated in FIG. 9. With reference to the preceding figures and associated description, the example machine readable instructions 900 begin with the example controller 116 determining equipment parameter(s) automatically or via manual entry. In some examples, the equipment parameter determiner 216 determines equipment parameter(s) automatically (e.g., based on data from the implement 108 and/or from the implement sensors 112 a-l), and/or via manual entry (e.g., via entry from the operator on the user interface 122).
At block 904, the example controller 116 begins manual or automated motion of the vehicle 102 across the field. In some examples, the operator controls motion of the vehicle 102 by commanding the drive commander 222, and controls usage of the implement 108 by commanding the implement commander 220. In some examples, the controller 116 automatically operates the vehicle via the drive commander 222 and the implement commander 220 by generating general guidance lines (e.g., evenly spaced, linear guidance lines) based on a field dimension and continually operating upon the field with all portions of the implement 108 via the implement commander 220.
At block 906, the example controller 116 collects location data during operation. In some examples, the location data analyzer 204 collects GPS data and inertial data from the location sensor 114 during operation of the vehicle 102.
At block 908, the example controller 116 determines if the field operation is complete. In some examples, the user command accessor 212 determines, based on user commands entered into the user interface 122, whether the field operation is complete. In response to the field operation being complete, processing transfers to block 910. Conversely, in response to the field operation not being complete, processing transfers to block 906.
At block 910, the example controller 116 generates a map for the field which has been operated upon. In some examples, the field map generator 206 generates a map for the field. For example, the field map generator 206 can generate the map based on location data from the location data analyzer 204 and/or implement operation data from the operation performance tracker 224.
At block 912, the example controller 116 stores the field map and/or uploads the field map to a network and/or central computing system. In some examples, the map data store 208 is used to store the field map.
Example machine readable instructions 100 that may be executed by the controller 116 to automatically guide a vehicle along guidance line(s) and perform an operation on a field are illustrated in FIG. 10. With reference to the preceding figures and associated description, the example machine readable instructions 1000 begin with the example controller 116 advancing the vehicle 102 along a guidance line accounting for curvature and/or contour of the field (Block 1002). In some examples, the drive commander 222 causes the vehicle 102 to advance along the guidance line through the field.
At block 1004, the example controller 116 activates portions of the implement 108 to perform operations on rows which have not been operated upon. In some examples, the implement commander 220 activates (e.g., enables, causes use of, etc.) portions of the implement 108 to perform operations on crop rows which have not yet been operated upon, based on implement operation data as analyzed by the operation performance tracker 224 and based on the field map, as analyzed by the field map analyzer 210.
At block 1006, the example controller 116 records implement operation data. In some examples, the operation performance tracker 224 records implement operation data by storing it in association with a field map of the current field being operated upon in the map data store 208.
FIG. 11 is a block diagram of an example processor platform 1100 structured to execute the instructions of FIGS. 8-10 to implement the controller 116 of FIGS. 1 and 2. The processor platform 1100 can be, for example, a server, a personal computer, a workstation, a self-learning machine (e.g., a neural network), a mobile device (e.g., a cell phone, a smart phone, a tablet such as an iPad™), a personal digital assistant (PDA), an Internet appliance, a DVD player, a CD player, a digital video recorder, a Blu-ray player, a gaming console, a personal video recorder, a set top box, a headset or other wearable device, or any other type of computing device.
The processor platform 1100 of the illustrated example includes a processor 1112. The processor 1112 of the illustrated example is hardware. For example, the processor 1112 can be implemented by one or more integrated circuits, logic circuits, microprocessors, GPUs, DSPs, or controllers from any desired family or manufacturer. The hardware processor may be a semiconductor based (e.g., silicon based) device. In this example, the processor implements the example field map handler 202, the example location data analyzer 204, the example field map generator 206, the example map data store 208, the example field map analyzer 210, the example user command accessor 212, the example user interface configurator 214, the example equipment parameter determiner 216, the example guidance line generator 218, the example implement commander 220, the example drive commander 222, and the example operation performance tracker 224.
The processor 1112 of the illustrated example includes a local memory 1113 (e.g., a cache). The processor 1112 of the illustrated example is in communication with a main memory including a volatile memory 1114 and a non-volatile memory 1116 via a bus 1118. The volatile memory 1114 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS® Dynamic Random Access Memory (RDRAM®) and/or any other type of random access memory device. The non-volatile memory 1116 may be implemented by flash memory and/or any other desired type of memory device. Access to the main memory 1114, 1116 is controlled by a memory controller.
The processor platform 1100 of the illustrated example also includes an interface circuit 1120. The interface circuit 1120 may be implemented by any type of interface standard, such as an Ethernet interface, a universal serial bus (USB), a Bluetooth® interface, a near field communication (NFC) interface, and/or a PCI express interface.
In the illustrated example, one or more input devices 1122 are connected to the interface circuit 1120. The input device(s) 1122 permit(s) a user to enter data and/or commands into the processor 1012. The input device(s) can be implemented by, for example, an audio sensor, a microphone, a camera (still or video), a keyboard, a button, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system.
One or more output devices 1124 are also connected to the interface circuit 1120 of the illustrated example. The output devices 1024 can be implemented, for example, by display devices (e.g., a light emitting diode (LED), an organic light emitting diode (OLED), a liquid crystal display (LCD), a cathode ray tube display (CRT), an in-place switching (IPS) display, a touchscreen, etc.), a tactile output device, a printer and/or speaker. The interface circuit 1120 of the illustrated example, thus, typically includes a graphics driver card, a graphics driver chip and/or a graphics driver processor.
The interface circuit 1120 of the illustrated example also includes a communication device such as a transmitter, a receiver, a transceiver, a modem, a residential gateway, a wireless access point, and/or a network interface to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 1126. The communication can be via, for example, an Ethernet connection, a digital subscriber line (DSL) connection, a telephone line connection, a coaxial cable system, a satellite system, a line-of-site wireless system, a cellular telephone system, etc.
The processor platform 1100 of the illustrated example also includes one or more mass storage devices 1128 for storing software and/or data Examples of such mass storage devices 1128 include floppy disk drives, hard drive disks, compact disk drives, Blu-ray disk drives, redundant array of independent disks (RAID) systems, and digital versatile disk (DVD) drives.
The machine executable instructions 1132 of FIGS. 1 and 2 may be stored in the mass storage device 1128, in the volatile memory 1114, in the non-volatile memory 1116, and/or on a removable non-transitory computer readable storage medium such as a CD or DVD.
From the foregoing, it will be appreciated that example methods, apparatus and articles of manufacture have been disclosed that enable accurate automated operation of an agricultural vehicle to perform operations on a field by utilizing field mapping data to generate guidance lines that account for differences between types of implements, unique topographical features of the field, unique crop row geometries, and/or other factors.
Although certain example methods, apparatus and articles of manufacture have been disclosed herein, the scope of coverage of this patent is not limited thereto. On the contrary, this patent covers all methods, apparatus and articles of manufacture fairly falling within the scope of the claims of this patent.

Claims (26)

What is claimed is:
1. An apparatus comprising:
a guidance line generator to generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows, the locations of the plurality of crop rows determined based on an implement path travelled by an implement of the vehicle during the first operation, the implement path different from a vehicle path of the vehicle during the first operation, the location data from an implement sensor mounted on the implement, and (2) the implement to perform the second operation on the field;
a drive commander to cause the vehicle to traverse the field along the guidance line; and
an implement commander to cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
2. The apparatus of claim 1, wherein the guidance line generator is to generate the guidance line based on an operational width of the implement.
3. The apparatus of claim 1, wherein the implement commander is to cause portions of the implement to perform the second operation upon crop rows which have not been operated upon.
4. The apparatus of claim 1, wherein the first operation is at least one of a planting operation or a strip till operation, and the second operation is an operation that follows the first operation.
5. The apparatus of claim 1, further including an operation performance tracker to store operation performance data, the operation performance data indicating locations at which the second operation was performed by the implement.
6. The apparatus of claim 1, wherein the guidance line is to account for at least one of a curvature or contour of the field.
7. The apparatus of claim 1, further including an equipment parameter determiner to determine an operational width of the implement based on a signal from the implement or a manual input from an operator into a user interface.
8. The apparatus of claim 1, further including a user interface configurator to cause a user interface in a cab of the vehicle to display a portion of the field map and the guidance line.
9. The apparatus of claim 1, wherein the field map is generated based on GPS data, inertial data, and implement sensor data during the first operation.
10. The apparatus of claim 1, wherein the guidance line is to prevent the second operation on portions of the field on which the second operation has been performed.
11. A computer readable storage medium comprising computer readable instructions that, when executed, cause a processor to:
generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows, the locations of the plurality of crop rows determined based on an implement path travelled by an implement of the vehicle during the first operation, the implement path different from a vehicle path of the vehicle during the first operation, the location data from an implement sensor mounted on the implement, and (2) the implement of the vehicle, the implement to perform the second operation on the field;
cause the vehicle to traverse the field along the guidance line; and
cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
12. The computer readable storage medium of claim 11, wherein generating the guidance line is based on an operational width of the implement.
13. The computer readable storage medium of claim 11, wherein the instructions, when executed, cause the processor to cause portions of the implement to perform the second operation upon crop rows which have not been operated upon.
14. The computer readable storage medium of claim 11, wherein the first operation is at least one of a planting operation or a strip till operation, and the second operation is an operation that follows the first operation.
15. The computer readable storage medium of claim 11, wherein the instructions, when executed, further cause the processor to store operation performance data, the operation performance data indicating locations at which the second operation was performed by the implement.
16. The computer readable storage medium of claim 11, wherein the guidance line is to account for at least one of a curvature or contour of the field.
17. A method comprising:
generating a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows, the locations of the plurality of crop rows determined based on an implement path travelled by an implement of the vehicle during the first operation, the implement path different from a vehicle path of the vehicle during the first operation, the location data from an implement sensor mounted on the implement, and (2) the implement to perform the second operation on the field;
causing the vehicle to traverse the field along the guidance line; and
causing the implement to perform the second operation as the vehicle traverses the field along the guidance line.
18. The method of claim 17, wherein generating the guidance line is based on an operational width of the implement.
19. The method of claim 17, further including causing portions of the implement to perform the second operation upon crop rows which have not been operated upon.
20. The method of claim 17, wherein the first operation is at least one of a planting operation or a strip till operation, and the second operation is at least one of a spraying operation, a harvesting operation, a fertilizing operation, a planting operation, a strip till operation, or a planting without till operation.
21. The method of claim 17, wherein the guidance line is to account for curvature of the field.
22. An apparatus comprising:
memory;
instructions in the memory; and
processor circuitry to execute the instructions to:
generate a guidance line for operation of a vehicle during a second operation on a field, the guidance line based on (1) a field map generated from location data collected during a first operation in the field, the field map including locations of a plurality of crop rows, the locations of the plurality of crop rows determined based on an implement path travelled by an implement of the vehicle during the first operation, the implement path different from a vehicle path of the vehicle during the first operation, the location data from an implement sensor mounted on the implement, and (2) the implement to perform the second operation on the field;
cause the vehicle to traverse the field along the guidance line; and
cause the implement to perform the second operation as the vehicle traverses the field along the guidance line.
23. The apparatus of claim 22, wherein the processor circuitry is to execute the instructions to generate the guidance line based on an operational width of the implement.
24. The apparatus of claim 22, wherein the processor circuitry is to execute the instructions to cause portions of the implement to perform the second operation upon crop rows which have not been operated upon.
25. The apparatus of claim 22, wherein the first operation is at least one of a planting operation or a strip till operation, and the second operation is at least one of a spraying operation, a harvesting operation, a fertilizing operation, a planting operation, a strip till operation, or a planting without till operation.
26. The apparatus of claim 22, wherein the guidance line is to account for curvature of the field.
US16/406,830 2019-05-08 2019-05-08 Apparatus and methods for field operations based on historical field operation data Active 2039-11-08 US11212954B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US16/406,830 US11212954B2 (en) 2019-05-08 2019-05-08 Apparatus and methods for field operations based on historical field operation data
EP20173207.0A EP3735811B1 (en) 2019-05-08 2020-05-06 Method for field operations based on historical field operation data
US17/645,264 US11825761B2 (en) 2019-05-08 2021-12-20 Apparatus and methods for field operations based on historical field operation data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/406,830 US11212954B2 (en) 2019-05-08 2019-05-08 Apparatus and methods for field operations based on historical field operation data

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/645,264 Continuation US11825761B2 (en) 2019-05-08 2021-12-20 Apparatus and methods for field operations based on historical field operation data

Publications (2)

Publication Number Publication Date
US20200352082A1 US20200352082A1 (en) 2020-11-12
US11212954B2 true US11212954B2 (en) 2022-01-04

Family

ID=70613612

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/406,830 Active 2039-11-08 US11212954B2 (en) 2019-05-08 2019-05-08 Apparatus and methods for field operations based on historical field operation data
US17/645,264 Active US11825761B2 (en) 2019-05-08 2021-12-20 Apparatus and methods for field operations based on historical field operation data

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/645,264 Active US11825761B2 (en) 2019-05-08 2021-12-20 Apparatus and methods for field operations based on historical field operation data

Country Status (2)

Country Link
US (2) US11212954B2 (en)
EP (1) EP3735811B1 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210003416A1 (en) * 2019-07-03 2021-01-07 Ag Leader Technology Apparatus, Systems And Methods For Automatic Steering Guidance And Visualization Of Guidance Paths
US11914379B2 (en) 2020-12-23 2024-02-27 Deere & Company Methods and apparatus to generate a path plan
JP7515437B2 (en) * 2021-04-14 2024-07-12 株式会社クボタ Driving Management System
JP7515436B2 (en) * 2021-04-14 2024-07-12 株式会社クボタ Route Generation System

Citations (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6236916B1 (en) 1999-03-29 2001-05-22 Caterpillar Inc. Autoguidance system and method for an agricultural machine
US20060178820A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for guiding an agricultural vehicle through a recorded template of guide paths
US20060178825A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for interactive selection of agricultural vehicle guide paths with varying curvature along their length
US20060178823A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for propagating agricultural vehicle guidance paths that have varying curvature along their length
US20060200294A1 (en) 2005-02-21 2006-09-07 Bernd Scheufler Electronic machine management system
US20070021913A1 (en) * 2003-03-20 2007-01-25 Satloc Llc Adaptive guidance system and method
US20070083299A1 (en) * 2005-10-12 2007-04-12 Lindores Robert J Method and system for determining the path of a mobile machine
US20080065286A1 (en) 2005-01-04 2008-03-13 Shufeng Han Method and system for guiding a vehicle with vision-based adjustment
US20080105177A1 (en) * 2006-11-03 2008-05-08 Dix Peter J Automatic path generation for tramlines
US20080249692A1 (en) * 2007-04-03 2008-10-09 Dix Peter J Method for creating end of row turns for agricultural vehicles
US20090037058A1 (en) 2007-07-31 2009-02-05 Aaron Matthew Senneff System and method for controlling a vehicle with a sequence of vehicle events
US20090118904A1 (en) * 2006-02-27 2009-05-07 Denis Allan Birnie Method and system for planning the path of an agricultural vehicle
US20100017075A1 (en) * 2006-03-10 2010-01-21 Norbert Beaujot Guiding Agricultural Implements
US8010261B2 (en) 2007-05-23 2011-08-30 Cnh America Llc Automatic steering correction of an agricultural harvester using integration of harvester header row sensors and harvester auto guidance system
US20120101634A1 (en) * 2010-10-25 2012-04-26 Lindores Robert J Crop treatment compatibility
US20120109614A1 (en) * 2010-10-25 2012-05-03 Lindores Robert J Crop characteristic estimation
US20140012732A1 (en) * 2010-10-25 2014-01-09 Trimble Navigation Limited Generating a crop recommendation
WO2014050524A1 (en) * 2012-09-26 2014-04-03 株式会社クボタ Agricultural management system and crop harvester
WO2014111387A1 (en) * 2013-01-18 2014-07-24 Naio Technologies Autonomous automated agricultural device
US20140236431A1 (en) 2013-02-20 2014-08-21 Deere And Company Soil compaction reduction system and method
US20140254861A1 (en) * 2013-03-08 2014-09-11 Raven Industries, Inc. Row guidance parameterization with hough transform
US20150321694A1 (en) * 2014-05-09 2015-11-12 Raven Industries, Inc. Refined row guidance parameterization with hough transform
US20160019560A1 (en) * 2014-07-16 2016-01-21 Raytheon Company Agricultural situational awareness tool
WO2016032901A1 (en) 2014-08-27 2016-03-03 Trimble Navigation Limited Three-dimensional elevation modeling for use in operating agricultural vehicles
US20160165795A1 (en) * 2014-12-15 2016-06-16 Irobot Corporation Robot lawnmower mapping
US20170090068A1 (en) * 2014-09-12 2017-03-30 The Climate Corporation Estimating soil properties within a field using hyperspectral remote sensing
DE102016216740A1 (en) * 2015-09-30 2017-03-30 Deere & Company System and method for using geo-fenced guidlines
WO2017061281A1 (en) * 2015-10-08 2017-04-13 ソニー株式会社 Information processing device and information processing method
WO2017066078A1 (en) * 2015-10-16 2017-04-20 The Climate Corporation A method for recommending seeding rate for corn seed using seed type and sowing row width
US20170127606A1 (en) * 2015-11-10 2017-05-11 Digi-Star, Llc Agricultural Drone for Use in Controlling the Direction of Tillage and Applying Matter to a Field
US20170169523A1 (en) * 2015-12-14 2017-06-15 The Climate Corporation Generating digital models of relative yield of a crop based on nitrate values in the soil
US20170270446A1 (en) * 2015-05-01 2017-09-21 360 Yield Center, Llc Agronomic systems, methods and apparatuses for determining yield limits
US20170293304A1 (en) * 2016-04-12 2017-10-12 Agjunction Llc Line acquisition path generation using curvature profiles
US20170300064A1 (en) * 2016-04-19 2017-10-19 CropZilla, Inc. Machine Guidance for Optimal Working Direction of Travel
US20180012347A1 (en) * 2016-07-11 2018-01-11 Harvest Moon Automation Inc. Inspecting plants for contamination
WO2018017746A1 (en) * 2016-07-22 2018-01-25 Precision Planting Llc Implements, systems, and methods for applying stress to early stage agricultural plants of agricultural fields
JP2018038291A (en) * 2016-09-05 2018-03-15 株式会社クボタ Work vehicle automatic traveling system and traveling route management device
WO2018101351A1 (en) * 2016-12-02 2018-06-07 株式会社クボタ Travel route management system and travel route determination device
US20180164117A1 (en) * 2016-12-14 2018-06-14 Kubota Corporation Travel route generation apparatus and method for generating travel route
CN108170084A (en) * 2016-12-07 2018-06-15 杨新高 A kind of agricultural equipment automatic control system and product
CN108170104A (en) * 2016-12-07 2018-06-15 杨新高 A kind of high intelligent agricultural machinery equipment and system
WO2018116770A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Work vehicle automatic traveling system
WO2018116772A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Work vehicle automatic traveling system
WO2018116769A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Traveling path management system
WO2018116771A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Traveling path determination device
EP3363273A1 (en) * 2017-02-16 2018-08-22 Amazonen-Werke H. Dreyer GmbH & Co. KG Agricultural machine system and method for planning lanes for processing an agricultural field
EP3400798A1 (en) * 2017-05-09 2018-11-14 CNH Industrial Italia S.p.A. An agricultural system
CA3065154A1 (en) * 2017-05-30 2018-12-06 Volta Robots S.R.L. Method for controlling a soil working means based on image processing and related system
US20180359904A1 (en) * 2017-06-19 2018-12-20 Cnh Industrial America Llc Path planning system for autonomous off-road vehicles
US10188029B1 (en) * 2014-10-20 2019-01-29 Hydro-Gear Limited Partnership Method of generating a three-dimensional map of a lawn and its use to improve mowing efficiency
CA3010410A1 (en) * 2017-08-31 2019-02-28 Cnh Industrial America Llc System and method for strip till implement guidance monitoring and adjustment
CA3010357A1 (en) * 2017-08-31 2019-02-28 Cnh Industrial America Llc Systems and method for determining trench closure by a planter or seeder
US20190170860A1 (en) * 2017-09-29 2019-06-06 Biotique Systems, Inc. System and method for automatically determining crop characteristics using unmanned aerial vehicle (UAV)
US20190230850A1 (en) * 2018-01-26 2019-08-01 Briggs & Stratton Corporation Systems and devices for autonomous lawn care
US20190259108A1 (en) * 2018-02-20 2019-08-22 Osram Gmbh Controlled Agricultural Systems and Methods of Managing Agricultural Systems
US20190289769A1 (en) * 2018-03-21 2019-09-26 Cnh Industrial America Llc Systems and methods for performing bale collection operations
US20190289779A1 (en) * 2016-07-14 2019-09-26 Precision Planting Llc Systems, implements, and methods for passive seed orientation within agricultural fields
US20190325534A1 (en) * 2018-04-24 2019-10-24 Indigo Ag, Inc. Agricultural transportation entity identification and selection
US20190369640A1 (en) * 2016-12-15 2019-12-05 Positec Power Tools (Suzhou) Co., Ltd Self-moving device, method for providing alarm about positioning fault in same, self-moving device, and automatic working system
WO2019245978A1 (en) * 2018-06-19 2019-12-26 Farm Jenny LLC Farm asset tracking, monitoring, and alerts
CA3105725A1 (en) * 2018-07-11 2020-01-16 Raven Industries, Inc. Agricultural control and interface system
US20200029487A1 (en) * 2018-07-30 2020-01-30 Cnh Industrial America Llc System and method for automatic implement depth measurement control
US20200037491A1 (en) * 2018-08-02 2020-02-06 Cnh Industrial America Llc System and method for monitoring field conditions of an adjacent swath within a field
US20200053962A1 (en) * 2018-08-20 2020-02-20 Cnh Industrial America Llc System and method for operating an agricultural harvester
US20200053961A1 (en) * 2018-08-20 2020-02-20 Cnh Industrial America Llc System and method for operating an agricultural harvester
DE202020000626U1 (en) * 2020-02-14 2020-05-07 Bernd Seidel Vibration-related identification, use and avoidance device for quantum entanglement
EP3729940A1 (en) * 2019-04-25 2020-10-28 CNH Industrial Sweden AB Method for adjusting the working depth of a plough implement and agricultural plough
US20200355667A1 (en) * 2019-05-10 2020-11-12 Cnh Industrial America Llc System and method for managing material accumulation relative to a ground engaging assembly of an agricultural implement based on the accumulation type
US20200383262A1 (en) * 2019-06-06 2020-12-10 Cnh Industrial America Llc System and method for controlling the operation of a seed-planting implement based on cover crop density
US20200383263A1 (en) * 2013-08-27 2020-12-10 Amvac Chemical Corporation System and method for treating individual seeds with liquid chemicals during the planting process
US20200390025A1 (en) * 2019-06-14 2020-12-17 Cnh Industrial America Llc System and method for monitoring soil conditions based on data received from a sensor mounted within a ground-engaging tool tooth
US20200390022A1 (en) * 2019-06-11 2020-12-17 Cnh Industrial America Llc System and method for controlling the operation of a seed-planting implement based on topographical features present within a field
US20210000006A1 (en) * 2019-07-02 2021-01-07 Bear Flag Robotics, Inc. Agricultural Lane Following
US20210000013A1 (en) * 2016-11-08 2021-01-07 Dogtooth Technologies Limited Robotic fruit picking system
US10936871B2 (en) * 2019-05-13 2021-03-02 Bao Tran Smart irrigation
US20210064050A1 (en) * 2019-08-30 2021-03-04 Deere & Company Methods and apparatus to record and execute mission plan
US20210146543A1 (en) * 2019-01-03 2021-05-20 Lucomm Technologies, Inc. Robotic Pallet
US11076589B1 (en) * 2020-10-16 2021-08-03 Verdant Robotics, Inc. Autonomous agricultural treatment system using map based targeting of agricultural objects
US20210265066A1 (en) * 2020-02-21 2021-08-26 Advanced Agrilytics Holdings, Llc Environmental matching techniques

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6522948B1 (en) * 2000-08-14 2003-02-18 Flexi-Coil Ltd. Agricultural product application tracking and control
US9213905B2 (en) * 2010-10-25 2015-12-15 Trimble Navigation Limited Automatic obstacle location mapping
US20230041214A1 (en) * 2013-02-04 2023-02-09 Dawn Equipment Company Agricultural systems
WO2015134886A1 (en) * 2014-03-06 2015-09-11 Raven Industries, Inc. System and method for sensing an edge
US9196100B1 (en) * 2014-06-16 2015-11-24 Deere & Company Equipment architecture for high definition data
AU2016274391B2 (en) * 2015-06-08 2020-08-13 Climate Llc Agricultural data analysis
US10719638B2 (en) * 2016-08-11 2020-07-21 The Climate Corporation Delineating management zones based on historical yield maps
US10238028B2 (en) * 2016-08-11 2019-03-26 The Climate Corporation Automatically detecting outlier values in harvested data
US10255670B1 (en) * 2017-01-08 2019-04-09 Dolly Y. Wu PLLC Image sensor and module for agricultural crop improvement
US10882065B2 (en) * 2017-02-28 2021-01-05 Deere & Company Agricultural vehicle with adjustable row unit
US10654063B2 (en) * 2017-02-28 2020-05-19 Deere & Company Adjustable row unit and agricultural vehicle with adjustable row unit
US10799903B2 (en) * 2017-02-28 2020-10-13 Deere & Company Adjustable row unit and vehicle with adjustable row unit
US10694734B2 (en) * 2017-02-28 2020-06-30 Deere & Company Adjustable row unit and sprayer vehicle with adjustable row unit
US10575460B2 (en) * 2017-02-28 2020-03-03 Deere & Company Adjustable row unit and vehicle with adjustable row unit
US10776733B2 (en) * 2017-06-26 2020-09-15 Acuitus Ag, Inc. Computer platform for controlling agricultural assets
US10795372B2 (en) * 2017-06-27 2020-10-06 Deere & Company Automatic end of row turning control system for a work vehicle by learning from operator
US10524407B2 (en) * 2017-06-30 2020-01-07 Deere & Company Crop row sensing on vehicle with multiple, independently steerable axles/wheels
US10438302B2 (en) * 2017-08-28 2019-10-08 The Climate Corporation Crop disease recognition and yield estimation
US10755367B2 (en) * 2018-05-10 2020-08-25 The Climate Corporation Analysis and presentation of agricultural data
EP3894851A4 (en) * 2018-12-10 2022-10-19 Climate LLC Mapping field anomalies using digital images and machine learning models
US11154004B2 (en) * 2019-04-26 2021-10-26 Deere & Company Agricultural seed sensing and control system
US11315258B1 (en) * 2019-08-19 2022-04-26 Ag Leader Technology, Inc. Optical system for tracking the heading and position of an implement compared to the pulling tractor and other uses
US11340092B2 (en) * 2020-01-29 2022-05-24 Deere & Company Work vehicle display systems and methods for automatic section control lookahead symbology
WO2021178998A1 (en) * 2020-03-02 2021-09-10 Raven Industries, Inc. Guidance systems and methods
WO2021207468A1 (en) * 2020-04-09 2021-10-14 Kinze Manufacturing, Inc. Simultaneous mapped view of historical and real-time geospatial and non-geospatial data
US20220226871A1 (en) * 2021-01-19 2022-07-21 Deere & Company Image-based roll gap optimization for roller conditioner systems
US12112546B2 (en) * 2021-04-30 2024-10-08 Deere & Company Vision guidance system using dynamic edge detection
US20230189688A1 (en) * 2021-12-22 2023-06-22 Agco Corporation Toolbar position mapping of an agricultural implement

Patent Citations (81)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6236916B1 (en) 1999-03-29 2001-05-22 Caterpillar Inc. Autoguidance system and method for an agricultural machine
US20070021913A1 (en) * 2003-03-20 2007-01-25 Satloc Llc Adaptive guidance system and method
US20080065286A1 (en) 2005-01-04 2008-03-13 Shufeng Han Method and system for guiding a vehicle with vision-based adjustment
US20060178820A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for guiding an agricultural vehicle through a recorded template of guide paths
US20060178825A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for interactive selection of agricultural vehicle guide paths with varying curvature along their length
US20060178823A1 (en) * 2005-02-04 2006-08-10 Novariant, Inc. System and method for propagating agricultural vehicle guidance paths that have varying curvature along their length
US20060200294A1 (en) 2005-02-21 2006-09-07 Bernd Scheufler Electronic machine management system
US20070083299A1 (en) * 2005-10-12 2007-04-12 Lindores Robert J Method and system for determining the path of a mobile machine
US20090118904A1 (en) * 2006-02-27 2009-05-07 Denis Allan Birnie Method and system for planning the path of an agricultural vehicle
US20100017075A1 (en) * 2006-03-10 2010-01-21 Norbert Beaujot Guiding Agricultural Implements
US20080105177A1 (en) * 2006-11-03 2008-05-08 Dix Peter J Automatic path generation for tramlines
US20080249692A1 (en) * 2007-04-03 2008-10-09 Dix Peter J Method for creating end of row turns for agricultural vehicles
US8010261B2 (en) 2007-05-23 2011-08-30 Cnh America Llc Automatic steering correction of an agricultural harvester using integration of harvester header row sensors and harvester auto guidance system
US20090037058A1 (en) 2007-07-31 2009-02-05 Aaron Matthew Senneff System and method for controlling a vehicle with a sequence of vehicle events
US20120101634A1 (en) * 2010-10-25 2012-04-26 Lindores Robert J Crop treatment compatibility
US20120109614A1 (en) * 2010-10-25 2012-05-03 Lindores Robert J Crop characteristic estimation
US20140012732A1 (en) * 2010-10-25 2014-01-09 Trimble Navigation Limited Generating a crop recommendation
WO2014050524A1 (en) * 2012-09-26 2014-04-03 株式会社クボタ Agricultural management system and crop harvester
WO2014111387A1 (en) * 2013-01-18 2014-07-24 Naio Technologies Autonomous automated agricultural device
US20140236431A1 (en) 2013-02-20 2014-08-21 Deere And Company Soil compaction reduction system and method
US20140254861A1 (en) * 2013-03-08 2014-09-11 Raven Industries, Inc. Row guidance parameterization with hough transform
US20200383263A1 (en) * 2013-08-27 2020-12-10 Amvac Chemical Corporation System and method for treating individual seeds with liquid chemicals during the planting process
US20150321694A1 (en) * 2014-05-09 2015-11-12 Raven Industries, Inc. Refined row guidance parameterization with hough transform
US20160019560A1 (en) * 2014-07-16 2016-01-21 Raytheon Company Agricultural situational awareness tool
WO2016032901A1 (en) 2014-08-27 2016-03-03 Trimble Navigation Limited Three-dimensional elevation modeling for use in operating agricultural vehicles
US20160057920A1 (en) * 2014-08-27 2016-03-03 Trimble Navigation Limited Three-Dimensional Elevation Modeling For Use In Operating Agricultural Vehicles
US20170090068A1 (en) * 2014-09-12 2017-03-30 The Climate Corporation Estimating soil properties within a field using hyperspectral remote sensing
US10188029B1 (en) * 2014-10-20 2019-01-29 Hydro-Gear Limited Partnership Method of generating a three-dimensional map of a lawn and its use to improve mowing efficiency
US20160165795A1 (en) * 2014-12-15 2016-06-16 Irobot Corporation Robot lawnmower mapping
US20170270446A1 (en) * 2015-05-01 2017-09-21 360 Yield Center, Llc Agronomic systems, methods and apparatuses for determining yield limits
DE102016216740A1 (en) * 2015-09-30 2017-03-30 Deere & Company System and method for using geo-fenced guidlines
WO2017061281A1 (en) * 2015-10-08 2017-04-13 ソニー株式会社 Information processing device and information processing method
WO2017066078A1 (en) * 2015-10-16 2017-04-20 The Climate Corporation A method for recommending seeding rate for corn seed using seed type and sowing row width
US20170127606A1 (en) * 2015-11-10 2017-05-11 Digi-Star, Llc Agricultural Drone for Use in Controlling the Direction of Tillage and Applying Matter to a Field
US20170169523A1 (en) * 2015-12-14 2017-06-15 The Climate Corporation Generating digital models of relative yield of a crop based on nitrate values in the soil
US20170293304A1 (en) * 2016-04-12 2017-10-12 Agjunction Llc Line acquisition path generation using curvature profiles
US20170300064A1 (en) * 2016-04-19 2017-10-19 CropZilla, Inc. Machine Guidance for Optimal Working Direction of Travel
US20180012347A1 (en) * 2016-07-11 2018-01-11 Harvest Moon Automation Inc. Inspecting plants for contamination
US20190289779A1 (en) * 2016-07-14 2019-09-26 Precision Planting Llc Systems, implements, and methods for passive seed orientation within agricultural fields
WO2018017746A1 (en) * 2016-07-22 2018-01-25 Precision Planting Llc Implements, systems, and methods for applying stress to early stage agricultural plants of agricultural fields
JP2018038291A (en) * 2016-09-05 2018-03-15 株式会社クボタ Work vehicle automatic traveling system and traveling route management device
US20210000013A1 (en) * 2016-11-08 2021-01-07 Dogtooth Technologies Limited Robotic fruit picking system
WO2018101351A1 (en) * 2016-12-02 2018-06-07 株式会社クボタ Travel route management system and travel route determination device
CN108170104A (en) * 2016-12-07 2018-06-15 杨新高 A kind of high intelligent agricultural machinery equipment and system
CN108170084A (en) * 2016-12-07 2018-06-15 杨新高 A kind of agricultural equipment automatic control system and product
US20180164117A1 (en) * 2016-12-14 2018-06-14 Kubota Corporation Travel route generation apparatus and method for generating travel route
US20190369640A1 (en) * 2016-12-15 2019-12-05 Positec Power Tools (Suzhou) Co., Ltd Self-moving device, method for providing alarm about positioning fault in same, self-moving device, and automatic working system
WO2018116770A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Work vehicle automatic traveling system
WO2018116771A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Traveling path determination device
WO2018116772A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Work vehicle automatic traveling system
WO2018116769A1 (en) * 2016-12-19 2018-06-28 株式会社クボタ Traveling path management system
EP3363273A1 (en) * 2017-02-16 2018-08-22 Amazonen-Werke H. Dreyer GmbH & Co. KG Agricultural machine system and method for planning lanes for processing an agricultural field
EP3400798A1 (en) * 2017-05-09 2018-11-14 CNH Industrial Italia S.p.A. An agricultural system
US20180325012A1 (en) * 2017-05-09 2018-11-15 Cnh Industrial America Llc Agricultural system
CA3065154A1 (en) * 2017-05-30 2018-12-06 Volta Robots S.R.L. Method for controlling a soil working means based on image processing and related system
US20180359904A1 (en) * 2017-06-19 2018-12-20 Cnh Industrial America Llc Path planning system for autonomous off-road vehicles
CA3010357A1 (en) * 2017-08-31 2019-02-28 Cnh Industrial America Llc Systems and method for determining trench closure by a planter or seeder
CA3010410A1 (en) * 2017-08-31 2019-02-28 Cnh Industrial America Llc System and method for strip till implement guidance monitoring and adjustment
US20190170860A1 (en) * 2017-09-29 2019-06-06 Biotique Systems, Inc. System and method for automatically determining crop characteristics using unmanned aerial vehicle (UAV)
US20190230850A1 (en) * 2018-01-26 2019-08-01 Briggs & Stratton Corporation Systems and devices for autonomous lawn care
US20190259108A1 (en) * 2018-02-20 2019-08-22 Osram Gmbh Controlled Agricultural Systems and Methods of Managing Agricultural Systems
US20190289769A1 (en) * 2018-03-21 2019-09-26 Cnh Industrial America Llc Systems and methods for performing bale collection operations
US20190325534A1 (en) * 2018-04-24 2019-10-24 Indigo Ag, Inc. Agricultural transportation entity identification and selection
WO2019245978A1 (en) * 2018-06-19 2019-12-26 Farm Jenny LLC Farm asset tracking, monitoring, and alerts
CA3105725A1 (en) * 2018-07-11 2020-01-16 Raven Industries, Inc. Agricultural control and interface system
US20200029487A1 (en) * 2018-07-30 2020-01-30 Cnh Industrial America Llc System and method for automatic implement depth measurement control
US20200037491A1 (en) * 2018-08-02 2020-02-06 Cnh Industrial America Llc System and method for monitoring field conditions of an adjacent swath within a field
US20200053962A1 (en) * 2018-08-20 2020-02-20 Cnh Industrial America Llc System and method for operating an agricultural harvester
US20200053961A1 (en) * 2018-08-20 2020-02-20 Cnh Industrial America Llc System and method for operating an agricultural harvester
US20210146543A1 (en) * 2019-01-03 2021-05-20 Lucomm Technologies, Inc. Robotic Pallet
EP3729940A1 (en) * 2019-04-25 2020-10-28 CNH Industrial Sweden AB Method for adjusting the working depth of a plough implement and agricultural plough
US20200355667A1 (en) * 2019-05-10 2020-11-12 Cnh Industrial America Llc System and method for managing material accumulation relative to a ground engaging assembly of an agricultural implement based on the accumulation type
US10936871B2 (en) * 2019-05-13 2021-03-02 Bao Tran Smart irrigation
US20200383262A1 (en) * 2019-06-06 2020-12-10 Cnh Industrial America Llc System and method for controlling the operation of a seed-planting implement based on cover crop density
US20200390022A1 (en) * 2019-06-11 2020-12-17 Cnh Industrial America Llc System and method for controlling the operation of a seed-planting implement based on topographical features present within a field
US20200390025A1 (en) * 2019-06-14 2020-12-17 Cnh Industrial America Llc System and method for monitoring soil conditions based on data received from a sensor mounted within a ground-engaging tool tooth
US20210000006A1 (en) * 2019-07-02 2021-01-07 Bear Flag Robotics, Inc. Agricultural Lane Following
US20210064050A1 (en) * 2019-08-30 2021-03-04 Deere & Company Methods and apparatus to record and execute mission plan
DE202020000626U1 (en) * 2020-02-14 2020-05-07 Bernd Seidel Vibration-related identification, use and avoidance device for quantum entanglement
US20210265066A1 (en) * 2020-02-21 2021-08-26 Advanced Agrilytics Holdings, Llc Environmental matching techniques
US11076589B1 (en) * 2020-10-16 2021-08-03 Verdant Robotics, Inc. Autonomous agricultural treatment system using map based targeting of agricultural objects

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
European Search Report issued in counterpart application No. 20173207.0 dated Oct. 5, 2020 (07 pages).

Also Published As

Publication number Publication date
US20220110240A1 (en) 2022-04-14
EP3735811A1 (en) 2020-11-11
US11825761B2 (en) 2023-11-28
EP3735811B1 (en) 2023-06-07
US20200352082A1 (en) 2020-11-12

Similar Documents

Publication Publication Date Title
US11825761B2 (en) Apparatus and methods for field operations based on historical field operation data
US11778936B2 (en) Apparatus and methods for vehicle steering to follow a curved path
US7591226B2 (en) Automatic path generation for tramlines
US7848865B2 (en) Method and system for applying materials to crops
Luck et al. Potential for pesticide and nutrient savings via map-based automatic boom section control of spray nozzles
US20210064050A1 (en) Methods and apparatus to record and execute mission plan
US20160057922A1 (en) System and method for controlling machinery for randomizing and replicating predetermined agronomic input levels
US20150305226A1 (en) System and method for sensor-based crop management
US11272701B2 (en) Method for remediating developmentally delayed plants
CA2968365A1 (en) System and methods for identifying fields and tasks
EP3361217A1 (en) Division map creation system and method for creating a division map
US8875641B2 (en) Agricultural implement pass marker system
JP2019088217A (en) Farm field work support terminal, farm field work machine and farm field work support program
BR102021026299A2 (en) METHOD FOR CONTROLLING AN AGRICULTURAL APPLICATION RATE, ONE OR MORE NON-TRANSITORY COMPUTER-READABLE MEDIA, AND APPARATUS FOR CONTROLLING AN AGRICULTURAL APPLICATION RATE
US12085955B2 (en) Methods, apparatus, and articles of manufacture to select track paths for one or more vehicles in a field
EP4091420A1 (en) Methods and apparatus to identify headlands
Tran et al. GNSS positioning and navigation-a foundational element of digital farming
JP2021023263A (en) Work field identification device
US20240192685A1 (en) Method and apparatus for zone mapping
US20230284548A1 (en) Method for performing driving route optimization
US20220128993A1 (en) Methods, apparatus, and articles of manufacture to display acquisition paths
Ortiz Rodríguez et al. COMPARISON OF PERFORMANCE IN MECHANISED AGRICULTURAL TASKS BETWEEN PRECISION AND CONVENTIONAL AGRICULTURE ON FARMS IN TOLIMA (COLOMBIA)?
WO2023170222A1 (en) Computer-implemented method for providing variable application rate data
Ćejvanović et al. Application of new technologies and precision agriculture.
ANUSHA et al. A New Approach for Exactitude Crop Growing Robot using Wireless Technology & LPC2148

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: DEERE & COMPANY, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, QIANG R.;MAEDER, CURTIS A.;PICKETT, TERENCE D.;AND OTHERS;REEL/FRAME:049131/0260

Effective date: 20190508

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE