Each view has its own coordinate system.
A coordinate system of any kind has a coordinate origin (0,0). Thus, the boundaries of any representation are the "boundaries" of this coordinate system, i.e. (0, 0, width, height).
A position of any kind is calculated based on the coordinate system of the parent view or superview . This is a frame of any kind, which therefore will be (x, y, width, height), where x and y represent the left and top or the position of the upper left corner in the original coordinate system.
The ios documentation explains this with this image

And for the latter, since AnchorPoint is โinsideโ the view, it is represented in the coordinate system of the view and therefore must use borders, as well as the center property.
This is also true for CALayer, where the boundaries of the layers are defined in their own coordinate system, and their frames are determined by their supervisors. There are many differences, some of them are direct, for example, layers have much deeper control, so you can set zordering and have a reference point for working with transformations, etc. Some of them are subtle, for example, for example, the frame CALayer property is not animated (i.e., its change in the animation block is not affected).
And I suggest you try smaller programs to understand the differences in addition to just trying to extract everything from a book or two, or even in this case even SO.